id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1366402594
Emit warning for deprecated fields in v1beta1 Changes [WIP]This commit adds warnings when a user creates a CRD with a deprecated field. Fixes: #5373 /kind feature Submitter Checklist As the author of this PR, please check off the items in this checklist: [x] Has Docs included if any changes are user facing [x] Has Tests included if any functionality added or changed [x] Follows the commit message standard [x] Meets the Tekton contributor standards (including functionality, content, code) [x] Has a kind label. You can add one by adding a comment on this PR that contains /kind <type>. Valid types are bug, cleanup, design, documentation, feature, flake, misc, question, tep [x] Release notes block below has been updated with any user facing changes (API changes, bug fixes, changes requiring upgrade notices or deprecation warnings) [x] Release notes contains the string "action required" if the change requires additional action from users switching to the new release Release Notes NONE /assign @lbernick
gharchive/pull-request
2022-09-08T14:02:50
2025-04-01T06:45:57.260272
{ "authors": [ "JeromeJu" ], "repo": "tektoncd/pipeline", "url": "https://github.com/tektoncd/pipeline/pull/5457", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
439176211
Fix TestStorageBucketPipelineRun to actually, y'know, run. Changes It was using the UpdateConfigMap function from Knative test, but that assumes that configMap.Data is already a non-nil map. Since in a normal test scenario with the default config-artifact-bucket.yaml in config/, the relevant configMap.Data is in fact nil, the test as written will just panic all the time. Mind you, the test will still fail, but that'll be due to #819. We haven't caught this, or #819, due to not running the bucket-related tests in CI! /assign @bobcatfish /assign @vdemeester /assign @dlorenc /assign @dwnusbaum Submitter Checklist These are the criteria that every PR should meet, please check them off as you review them: [x] Includes tests (if functionality changed/added) [x] Includes docs (if user facing) [x] Commit messages follow commit message best practices See the contribution guide for more details. Release Notes n/a @abayer just checking my understanding: the test still won't actually run in CI, but I assume you verified this fix manually? Yup, in order to reproduce #819.
gharchive/pull-request
2019-05-01T14:06:45
2025-04-01T06:45:57.265851
{ "authors": [ "abayer", "bobcatfish" ], "repo": "tektoncd/pipeline", "url": "https://github.com/tektoncd/pipeline/pull/820", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1670957368
fully functional send page Fixes #223 also Fixes #174 presented in this draft PR also Fixes #253 solved in this PR currently merged Description In order to fulfill the requirement of having a fully functional "send page", three largely independent parts needed to be completed: (#253) The implementation of new notification banners for success and error messages. (#174) The visual design of the send page, including the HTML and CSS. (#223) Adding support for token sending in the new library, which can be further divided into three parts: Retrieving balances for each ERC20 or native token. Obtaining the gas price and estimating the cost of execution. Actual sending of the ERC20 or native token. Test scenarios goto this link login on EVM using an account with at least 1 liquid TLOS and a positive balance of any other ERC20 token goto Send Page you should see the send tokens form TLOS should be selected by default the estimated fee should be low but not zero (in TLOS) view token link should not be visible the whole balance should be set in the amount's input. the input should show the amount followed by a semi-transparent symbol and the end the symbol should not affect the input content and should not be editable or even selected the available amount should show the total token balance the fiat balance should be shown in the footer of the input field clean the amount field and press the available amount link the whole balance should be set in the amount's input. click the fiat balance the footer balance and the input balance should be interchanged the available amount should show the total fiat balance. the input should show the USD symbol and then the input field amount (not editable or selectable) change the token the view token contract link should be shown in the token selector footer the available balance should show the total balance for the new token selected if no price is available for that token, then the footers for the amount input are not shown. the estimated fee should be higher than the system token sending gas prepare an invalid form (no address, not enough balance, or non-positive amount) the send button should be disabled prepare a valid form using any token and press send The Metamask should present the sending confirmation with the correct information, confirm!! If no errors a success notification should be shown for at least 5 sec if you press the see transaction link you should be redirected to Teloscan to see the transaction if an error occurs you should see an error notification with the error info. Checklist: [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have cleaned up the code in the areas my change touches [x] My changes generate no new warnings [x] Any dependent changes have been merged and published in downstream modules [x] I have checked my code and corrected any misspellings [x] I have removed any unnecessary console messages [x] I have included all english text to the translation file [x] I have created a new issue with the required translations for the currently supported languages [ ] I have added appropriate test coverage symbol is being shown twice fiat value should be 2 decimal places there should be an error when the user has entered too many decimal places for the token i think. in this case pressing send doesnt work, but there is no feedback to the user that it's because there are too many decimal places at mobile the layout breaks. lets move the token picker to a new line on on XS devices. this will mean the available text label will need to be moved in the markup too it seems also there should be some validation on the address field. you can use this regex to check if the input has a valid ethereum address: const addressRegex = /^0x\d{40}$/; const isValid = addressRegex.test(address); if the user does not enter a valid regex, there should be some error text in my opinion. you may also want to add placeholder text like :placeholder=`e.g. 0x${'0'.repeat(40)}` sending fails if i try to send STLOS :( there is also a huge error message that takes up the whole screen (cant get a picture because it goes too fast), but basically it fails because it cannot estimate gas. error starts with: nnot estimate gas; transaction may fail or may require manual gas limit [ See: https://links.ethers.org/v5-errors-UNPREDICTABLE_GAS_LIMIT ] i think maybe we don't want to show the user the actual error message, maybe either show something generic and they can check the console if they want, or cut off the message after a certain number of characters I am still seeing an error when trying to send STLOS. printing the entire error leads to this: maybe we can parse the error message to get an error code, then display a message like "if error code = -32603, error mesage = "failed to estimate gas"' note: this error is actually handled in teloscan, it may help you debug: https://github.com/telosnetwork/teloscan/blob/dev/src/components/LoginModal.vue#L278 @Viterbo double 'back' button
gharchive/pull-request
2023-04-17T11:22:09
2025-04-01T06:45:57.411116
{ "authors": [ "EJSG", "Viterbo", "donnyquixotic" ], "repo": "telosnetwork/telos-wallet", "url": "https://github.com/telosnetwork/telos-wallet/pull/257", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1864045457
🛑 Secret Site2 is down In 5c0479f, Secret Site2 ($SECRET_SITE_2) was down: HTTP code: 404 Response time: 80 ms Resolved: Secret Site2 is back up in 4d4ef0f after 59 days, 14 hours, 42 minutes.
gharchive/issue
2023-08-23T21:33:54
2025-04-01T06:45:57.426952
{ "authors": [ "templain" ], "repo": "templain/mywatcher", "url": "https://github.com/templain/mywatcher/issues/1327", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2050064918
🛑 QRCODE VIEWER is down In b1b8d9a, QRCODE VIEWER (https://qrcode-viewer.onrender.com/qrcode-viewer/) was down: HTTP code: 0 Response time: 0 ms Resolved: QRCODE VIEWER is back up in 1681703 after 9 minutes.
gharchive/issue
2023-12-20T08:16:23
2025-04-01T06:45:57.429456
{ "authors": [ "templain" ], "repo": "templain/mywatcher", "url": "https://github.com/templain/mywatcher/issues/3024", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1825998752
🛑 Secret Site2 is down In 84eec6c, Secret Site2 ($SECRET_SITE_2) was down: HTTP code: 404 Response time: 84 ms Resolved: Secret Site2 is back up in d60ca94.
gharchive/issue
2023-07-28T08:50:07
2025-04-01T06:45:57.431612
{ "authors": [ "templain" ], "repo": "templain/mywatcher", "url": "https://github.com/templain/mywatcher/issues/645", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
300165685
部分安卓手机uploadFile方法直接返回{"readyState":0,"responseText":"","status":0,"statusText":"error"} 部分安卓手机uploadFile方法直接返回{"readyState":0,"responseText":"","status":0,"statusText":"error"}, uploading... curr progress is NaN 如果是部分机型出现的错误,可以尝试配置 CORS AllowHeaders: * 如果还是不行,提供一下请求返回的 RequestId AllowHeaders已配置为*,目前只发现oppo r9跟锤子有问题,在successCallback返回的只有{"readyState":0,"responseText":"","status":0,"statusText":"error"} 返回的 header 里有个 x-cos-request-id 提供一下? { "code": -121, "message": "ERROR_CMD_CORSRULE_NOT_MATCH", "request_id": "NWE5M2ViODhfZDgyNzVkNjRfZDk1OF8yOTA5OTc=" } ERROR_CMD_CORSRULE_NOT_MATCH 这个就是 CORS 配置有误 请检查请求参数是不是搞错了,导致请求到别的 bucket 去,然后检查你 bucket 的 CORS 是否正确。 如果还不行发一下你的 appid 和 bucket 看看 这个是公用的方法,其他机型都能正常上传,只有oppo r9 跟锤子会报这个错误,bucket确认是正确的 一般这个表现是因为某些浏览器请求的时候会带有特定的 Access-Control-Request-Headers: origin,accept,content-type 你再次确认下 CORS 配置:AllowHeaders: * 如果还不行,请贴上 appid 和 bucket 我检查下,或者加我 QQ:459452372 抱歉,刚去检查了下配置,AllowHeaders没有配上导致的,耽误您时间了,实在抱歉,现在好了 ok
gharchive/issue
2018-02-26T09:35:05
2025-04-01T06:45:57.486373
{ "authors": [ "carsonxu", "warrior-bing" ], "repo": "tencentyun/cos-js-sdk-v4", "url": "https://github.com/tencentyun/cos-js-sdk-v4/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1141512594
Do we support tf.nn.bidirectional_dynamic_rnn ? Hi: I only see CudnnRNN ,the project has not testcase for it,I not know if we support tf.nn.bidirectional_dynamic_rnn or normal simple rnn? thank you I don't think we do support out-of-the-box RNN ops, most in TensorFlow has been also deprecated in favor to Keras RNN cells and layers, so I guess this support will also as we port the Keras API into our framework. I don't think we do support out-of-the-box RNN ops, most in TensorFlow has been also deprecated in favor to Keras RNN cells and layers, so I guess this support will also as we port the Keras API into our framework. wow, when this framework will wrapper keras api? This is a own project on its own, leaded by @JimClarke5 , a lot of work as already been done by porting the optimizers, metrics, losses, regularizers and activations of Keras to Java. The plumbing for layers and models are coming on soon and from there it would be interesting to tackle the support for RNNs. If you are interested to know more about it @mullerhai and if you can, I suggest that you join the next SIG meeting where we discuss various topics related to TF Java.
gharchive/issue
2022-02-17T15:45:29
2025-04-01T06:45:57.603213
{ "authors": [ "karllessard", "mullerhai" ], "repo": "tensorflow/java", "url": "https://github.com/tensorflow/java/issues/421", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
352573335
padding='causal' is in document but cannot be used with tf.keras.layers.Conv1D System information Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Yes OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 16.04 Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: TensorFlow installed from (source or binary): I use docker tensorflow/tensorflow:1.10.0-devel-gpu-py3 TensorFlow version (use command below): 1.10.0 Python version: Python 3.5.2 Bazel version (if compiling from source): GCC/Compiler version (if compiling from source): CUDA/cuDNN version: GPU model and memory: Exact command to reproduce: see my gist below Describe the problem In the document, we can use causal option for padding. But causal option causes error in r1.10. We should remove causal from the document or support causal. There are related issues, #14933, #15000 and #15037. But all of them are closed. I think they should be reopen. Source code / logs https://gist.github.com/dhgrs/ca552f5804ddfb4db9669f9189eba76f Nagging Assignee @drpngx: It has been 14 days with no activity and this issue has an assignee. Please update the label and/or status accordingly. Nagging Assignee @drpngx: It has been 29 days with no activity and this issue has an assignee. Please update the label and/or status accordingly. This is fixed now in the latest nightly version, please pip install tf-nightly (or pip install tf-nightly-gpu for GPU support)
gharchive/issue
2018-08-21T14:39:34
2025-04-01T06:45:57.711287
{ "authors": [ "dhgrs", "omalleyt12", "tensorflowbutler" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/21763", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
384407988
Save/Load problem with keras.layers.ReLU Running this example x = tf.keras.layers.Input([1]) y = tf.keras.layers.ReLU()(x) mdl = tf.keras.models.Model(x, y) tf.keras.models.save_model(mdl, "./keras_model_tmp.hdf5", overwrite=True) tf.keras.models.load_model("./keras_model_tmp.hdf5") returns trackback /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/engine/saving.pyc in load_model(filepath, custom_objects, compile) 228 raise ValueError('No model found in config file.') 229 model_config = json.loads(model_config.decode('utf-8')) --> 230 model = model_from_config(model_config, custom_objects=custom_objects) 231 232 # set weights /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/engine/saving.pyc in model_from_config(config, custom_objects) 308 '`Sequential.from_config(config)`?') 309 from tensorflow.python.keras.layers import deserialize # pylint: disable=g-import-not-at-top --> 310 return deserialize(config, custom_objects=custom_objects) 311 312 /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/layers/serialization.pyc in deserialize(config, custom_objects) 62 module_objects=globs, 63 custom_objects=custom_objects, ---> 64 printable_module_name='layer') /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/utils/generic_utils.pyc in deserialize_keras_object(identifier, module_objects, custom_objects, printable_module_name) 171 custom_objects=dict( 172 list(_GLOBAL_CUSTOM_OBJECTS.items()) + --> 173 list(custom_objects.items()))) 174 with CustomObjectScope(custom_objects): 175 return cls.from_config(config['config']) /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/engine/network.pyc in from_config(cls, config, custom_objects) 1291 for layer_data in config['layers']: 1292 process_layer(layer_data) -> 1293 # Then we process nodes in order of layer depth. 1294 # Nodes that cannot yet be processed (if the inbound node 1295 # does not yet exist) are re-enqueued, and the process /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/engine/network.pyc in process_layer(layer_data) 1276 from tensorflow.python.keras.layers import deserialize as deserialize_layer # pylint: disable=g-import-not-at-top 1277 -> 1278 layer = deserialize_layer(layer_data, custom_objects=custom_objects) 1279 created_layers[layer_name] = layer 1280 /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/layers/serialization.pyc in deserialize(config, custom_objects) 62 module_objects=globs, 63 custom_objects=custom_objects, ---> 64 printable_module_name='layer') /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/utils/generic_utils.pyc in deserialize_keras_object(identifier, module_objects, custom_objects, printable_module_name) 173 list(custom_objects.items()))) 174 with CustomObjectScope(custom_objects): --> 175 return cls.from_config(config['config']) 176 else: 177 # Then `cls` may be a function returning a class. /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/engine/base_layer.pyc in from_config(cls, config) 1604 A layer instance. 1605 """ -> 1606 return cls(**config) 1607 1608 /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/layers/advanced_activations.py in __init__(self, max_value, negative_slope, threshold, **kwargs) 319 self.max_value = max_value 320 self.negative_slope = K.cast_to_floatx(negative_slope) --> 321 self.threshold = K.cast_to_floatx(threshold) 322 323 def call(self, inputs): /data/virtualenv/venv2/lib/python2.7/site-packages/tensorflow/python/keras/backend.pyc in cast_to_floatx(x) 231 ``` 232 """ --> 233 return np.asarray(x, dtype=_FLOATX) 234 235 /data/virtualenv/venv2/lib/python2.7/site-packages/numpy/core/numeric.pyc in asarray(a, dtype, order) 499 500 """ --> 501 return array(a, dtype, copy=False, order=order) 502 503 TypeError: float() argument must be a string or a number I temporally fixed it in tensorflow/python/keras/layers/advanced_activations.py by changing get_config from def get_config(self): config = { 'max_value': self.max_value, 'negative_slope': self.negative_slope, 'threshold': self.threshold } base_config = super(ReLU, self).get_config() return dict(list(base_config.items()) + list(config.items())) to def get_config(self): config = { 'max_value': self.max_value, 'negative_slope': float(self.negative_slope), 'threshold': float(self.threshold) } base_config = super(ReLU, self).get_config() return dict(list(base_config.items()) + list(config.items())) However, the problem might be somewhere in deserialize_layer, as for a some reason negative_slope is a dictionary, not a numpy.ndarray @cataclysmus I am able to reproduce with tf 1.12.0, but not with tf-nightly. Can you try with tf-nightly and see if the issue still exist? Thanks @yongtang for taking a look. Previously I failed to check with nightly version. However I checked with TF nightly version : '1.13.0-dev20181108' now and the issue seems to be fixed. Will wait for @cataclysmus confirmation. @yongtang @ymodak Checked. No bug at the build 1.13.0-dev20181128 Are you satisfied with the resolution of your issue? Yes No
gharchive/issue
2018-11-26T16:24:54
2025-04-01T06:45:57.717122
{ "authors": [ "cataclysmus", "ymodak", "yongtang" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/23979", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
404187660
tf.contrib.opt.ScipyOptimizerInterface error Hello. I tried to use tf.contrib.opt.ScipyOptimizerInterface with an example code from your documentation. vector = tf.Variable([7., 7.], 'vector') loss = tf.reduce_sum(tf.square(vector)) optimizer = tf.contrib.opt.ScipyOptimizerInterface(loss, options={'maxiter': 100}) with tf.Session() as session: ---->optimizer.minimize(session). I am getting the following error: Tensors in list passed to 'values' of 'ConcatV2' Op have types [int32, float32, float32, float32, float32, float32, float32, float32, float32, float32, int32, float32, float32, float32, float32, int32, int32, float32, float32, float32, float32] that don't all match. OS Platform and Distribution: Ubuntu 18.04.1 TensorFlow installed from: source TensorFlow version: 1.12.0 Python version: 3.6 Bazel version: 0.21.0 GCC version: 7.3.0 CUDA and GPU: no running on GPU, because I don't have GPU. Thank you! ScipyOptimizer is moving from contrib to tensorflow/addons in #3 . I'm afraid that it would be a while, at least, before we could look into the issue. cc @seanpmorgan @facaiy is correct it will take some time before we're in a position to accept questions for addons ExternalOptimizers. As part of the moving and upgrading code from contrib to addons, we also need to step through the code and make sure we understand exactly how it works. In the mean time, I'll try to some troubleshooting -- but can not give any concrete time estimates. Contrib has been depreciated here in Tensorflow repo and moved to tensorflow/addons , please reopen the issue in https://github.com/tensorflow/addons/issues if this exists in latest version. Thank you
gharchive/issue
2019-01-29T09:07:32
2025-04-01T06:45:57.722560
{ "authors": [ "ciprianmihai94", "facaiy", "rthadur", "seanpmorgan" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/25279", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
419645948
tensorflow lite: error while converting frozen model to lite format System information Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Win 10 pro Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: NA TensorFlow installed from (source or binary): Using Anaconda navigator TensorFlow version (use command below): v1.12.0-9901-gf380d8b8e5 1.14.1-dev20190309 Python version: Python 3.6.8 Bazel version (if compiling from source): NA GCC/Compiler version (if compiling from source): NA CUDA/cuDNN version: using cpu version GPU model and memory: na You can collect some of this information using our environment capture script You can also obtain the TensorFlow version with python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)" Describe the current behavior I am trying to quantize a model that is made using tfslim library. I froze it using graph_util.convert_variables_to_constants function and tried to convert it to tflite formate for quantization. I am using win 10 and tf-nighlty. the following error appears: TOCO failed. See console for info. 2019-03-12 00:20:27.142300: I tensorflow/lite/toco/graph_transformations/graph_transformations.cc:39] Before Removing unused ops: 221 operators, 366 arrays (0 quantized) 2019-03-12 00:20:27.147170: I tensorflow/lite/toco/graph_transformations/graph_transformations.cc:39] Before general graph transformations: 221 operators, 366 arrays (0 quantized) 2019-03-12 00:20:27.243872: F tensorflow/lite/toco/graph_transformations/propagate_fixed_sizes.cc:722] Check failed: start_array.data_type == ArrayDataType::kInt32 Range op inputs must be int32. Fatal Python error: Aborted Current thread 0x00001f44 (most recent call first): File "c:\anaconda3\lib\site-packages\tensorflow\lite\toco\python\toco_from_protos.py", line 33 in execute File "c:\anaconda3\lib\site-packages\absl\app.py", line 251 in run_main File "c:\anaconda3\lib\site-packages\absl\app.py", line 300 in run File "c:\anaconda3\lib\site-packages\tensorflow\python\platform\app.py", line 40 in run File "c:\anaconda3\lib\site-packages\tensorflow\lite\toco\python\toco_from_protos.py", line 59 in main File "C:\anaconda3\Scripts\toco_from_protos.exe_main.py", line 9 in File "c:\anaconda3\lib\runpy.py", line 85 in _run_code File "c:\anaconda3\lib\runpy.py", line 193 in _run_module_as_main Describe the expected behavior Should convert frozen pb file to tflite with int8 quantization Code to reproduce the issue Provide a reproducible test case that is the bare minimum necessary to generate the problem. pb file in the below link https://drive.google.com/open?id=1Ba6aEuv2VTGD_bct24UBiMh7MFe_jkOr graph_def_file = "frozen_model_yolov3-tiny.pb" input_arrays = ["input_to_model"] output_arrays = ["concat_1"] converter = tf.lite.TFLiteConverter.from_frozen_graph( graph_def_file, input_arrays, output_arrays) tflite_model = converter.convert() open("converted_tiny_model.tflite", "wb").write(tflite_model) Other info / logs Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Thank you for your post. We noticed you have not filled out the fields in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Along with the template, please provide as many details as possible to find the root cause of the issue. Thank you for your post. We noticed you have not filled out the fields in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Along with the template, please provide as many details as possible to find the root cause of the issue. Ive updated the post. please have a look Thank you for your post. We noticed you have not filled out the fields in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Along with the template, please provide as many details as possible to find the root cause of the issue. Ive updated the post. please have a look @haozha111 Could you help to take a look at the model? I think this requires broadening supported types of Range, in both converter and runtime I downloaded your model, but I can't reproduce your issue. I used virtualenv and tf-nightly. (tensorflow_venv) haoliang@haoliang0:~/tensorflow_venv/bin$ python test.py 2019-03-19 16:37:17.875562: I tensorflow/core/platform/cpu_feature_guard.cc:142] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 FMA 2019-03-19 16:37:17.905232: I tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency: 3492135000 Hz 2019-03-19 16:37:17.906497: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x55a21cc6c070 executing computations on platform Host. Devices: 2019-03-19 16:37:17.906550: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): , 2019-03-19 16:37:18.050999: I tensorflow/core/grappler/devices.cc:53] Number of eligible GPUs (core count >= 8): 0 (Note: TensorFlow was not compiled with CUDA support) 2019-03-19 16:37:18.051125: I tensorflow/core/grappler/clusters/single_machine.cc:359] Starting new session 2019-03-19 16:37:18.283008: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:666] Optimization results for grappler item: graph_to_optimize 2019-03-19 16:37:18.283043: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] model_pruner: Graph size after: 236 nodes (-63), 253 edges (-63), time = 1.355ms. 2019-03-19 16:37:18.283054: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] implementation_selector: Graph size after: 236 nodes (0), 253 edges (0), time = 0.397ms. 2019-03-19 16:37:18.283071: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] function_optimizer: Graph size after: 236 nodes (0), 253 edges (0), time = 0.306ms. 2019-03-19 16:37:18.283078: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] constant folding: Graph size after: 158 nodes (-78), 173 edges (-80), time = 25.045ms. 2019-03-19 16:37:18.283084: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] shape_optimizer: Graph size after: 158 nodes (0), 173 edges (0), time = 0.505ms. 2019-03-19 16:37:18.283090: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] arithmetic_optimizer: Graph size after: 152 nodes (-6), 175 edges (2), time = 100.996ms. 2019-03-19 16:37:18.283096: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] loop_optimizer: Graph size after: 152 nodes (0), 175 edges (0), time = 0.661ms. 2019-03-19 16:37:18.283102: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] dependency_optimizer: Graph size after: 151 nodes (-1), 173 edges (-2), time = 1.329ms. 2019-03-19 16:37:18.283110: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] layout: Graph size after: 151 nodes (0), 173 edges (0), time = 0.286ms. 2019-03-19 16:37:18.283119: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] memory_optimizer: Graph size after: 151 nodes (0), 173 edges (0), time = 9.367ms. 2019-03-19 16:37:18.283125: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] model_pruner: Graph size after: 151 nodes (0), 173 edges (0), time = 0.983ms. 2019-03-19 16:37:18.283130: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] implementation_selector: Graph size after: 151 nodes (0), 173 edges (0), time = 0.257ms. 2019-03-19 16:37:18.283135: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] function_optimizer: Graph size after: 151 nodes (0), 173 edges (0), time = 0.254ms. 2019-03-19 16:37:18.283140: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] constant folding: Graph size after: 149 nodes (-2), 171 edges (-2), time = 12.494ms. 2019-03-19 16:37:18.283146: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] shape_optimizer: Graph size after: 149 nodes (0), 171 edges (0), time = 0.503ms. 2019-03-19 16:37:18.283152: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] arithmetic_optimizer: Graph size after: 149 nodes (0), 171 edges (0), time = 37.109ms. 2019-03-19 16:37:18.283160: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:668] dependency_optimizer: Graph size after: 149 nodes (0), 171 edges (0), time = 1.423ms. Could you check what version of Tensorflow you are using? @haozha111 will this snippet trigger the error? graph_def_file = "frozen_model_yolov3-tiny.pb" input_arrays = ["input_to_model"] output_arrays = ["concat_1"] converter = tf.lite.TFLiteConverter.from_frozen_graph(graph_def_file, input_arrays, output_arrays) converter.inference_type = tf.lite.constants.QUANTIZED_UINT8 input_arrays = converter.get_input_arrays() converter.quantized_input_stats = {input_arrays[0] : (0., 1.)} # mean, std_dev tflite_model = converter.convert() open("converted_tiny_model.tflite", "wb").write(tflite_model) I tried the same, unfortunately got exactly the same error. I'm using virtualenv and tf-nightly ('1.14.1-dev20190410') Thanks Mark. Now I can reproduce that exact same error. I think this will only fail when quantization is enabled. See the comments here[1], it looks like range op currently doesn't support quantization. So the check will fail when the input is not int32 or float32. I will transfer this issue to Pulkit since he is planning to support quantization for range op. [1]https://github.com/tensorflow/tensorflow/blob/c238fc4a25c71c71e1bfdbb6eac750c64aed52ad/tensorflow/lite/kernels/range.cc#L89 Thanks Mark. Now I can reproduce that exact same error. I think this will only fail when quantization is enabled. See the comments here[1], it looks like range op currently doesn't support quantization. So the check will fail when the input is not int32 or float32. I will transfer this issue to Pulkit since he is planning to support quantization for range op. [1]https://github.com/tensorflow/tensorflow/blob/c238fc4a25c71c71e1bfdbb6eac750c64aed52ad/tensorflow/lite/kernels/range.cc#L89 @gjraza did you solve this ? @nutsiepully @haozha111 , When will you support quantization for range op
gharchive/issue
2019-03-11T19:33:27
2025-04-01T06:45:57.753893
{ "authors": [ "Rachel07", "devimonica", "gjraza", "haozha111", "jvishnuvardhan", "marksunpeng", "miaout17" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/26586", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
436361739
Windows Build working with cuda 10.1 and vs2019 I am only writing this as an issue as I got tensorflow with win 10 building with cuda 10.1 and vs2019 . I am not sure if my changes can be general, they seem to work for me and I can train MNIST on GPU. The cuda changes could be relevant, the others might be working only for my system config. Please make sure that this is a build/installation issue. As per our GitHub Policy, we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template System information OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10 Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: TensorFlow installed from (source or binary): source TensorFlow version: 2.0a Python version: 3.6.5 Installed using virtualenv? pip? conda?: pip Bazel version (if compiling from source): 20 GCC/Compiler version (if compiling from source): visual studo 2019 CUDA/cuDNN version: 10.1/7 GPU model and memory: gtx 1080 ti , 4gv Describe the problem Cannot compile tensorflow 2.0 alpha on windows with cuda 10.1, and visual studio 2019 Provide the exact sequence of commands / steps that you executed before running into the problem Hardlinked visual studio 2019 as 2017 with mklink modified bazel scripts and stream operators as in : https://github.com/atemysemicolon/tensorflow/commit/54faf06d5c10e8875b9ce36b941004aceb44c8b9 Compiles Any other info / logs Visual 2019 does not seem to be supported by older versions of Bazel. Bazel 24.1 causes compilation issues using nvcc. However, cheating bazel into thinking 2019 is 2017 is good enough - by hardlinking the 2019 folder to 2017 in C:\Program Files\Microsoft Visual Studio.. I had to modify bazel to not generate workspaces for mac and linux, The most important one seems to be the modification in cuda_10_0.inc in tensorflow/stream_executer/cuda/... TensorFlow supports CUDA 10.0 (TensorFlow >= 1.13.0). Please refer this link and let us know in case of any clarification. Thanks! This guide is for 10.1 , not 10 Unfortunately, none of the TensorFlow versions officially support CUDA 10.1. As of now we have to build TensorFlow version>=1.13.0 using CUDA 10.0. Thanks! Automatically closing due to lack of recent activity. Please update the issue when new information becomes available, and we will reopen the issue. Thanks! I thought only when someone builds it , it could be officially supported xd
gharchive/issue
2019-04-23T19:54:28
2025-04-01T06:45:57.763251
{ "authors": [ "achandraa", "atemysemicolon" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/28086", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
461299150
ERROR: Cannot uninstall 'wrapt'. during upgrade System information OS Platform and Distribution (e.g., Linux Ubuntu 19.04 on Sony notebook): TensorFlow version: python Python 3.6.3 |Anaconda, Inc.| (default, Oct 13 2017, 12:02:49) [GCC 7.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tensorflow as tf /home/ronald/anaconda3/lib/python3.6/site-packages/h5py/__init__.py:34: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`. from ._conv import register_converters as _register_converters >>> print(tf.__version__) 2.0.0-alpha0 Describe the problem I try to upgrade to beta1 with: install tensorflow==2.0.0-beta1 at the end I get: Installing collected packages: wrapt, tensorflow Found existing installation: wrapt 1.10.11 ERROR: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. How to fix it? I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 It didn't work for me. Still getting that error. it work for me. Thank you I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 It work for me. Thank you Works for me, thanks! It didn't work for me. Still getting that error. I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 It worked for me also. conda update --all pip install tensorflow It didn't work for me. Still getting that error. try conda update wrapt conda update --all pip install tensorflow this worked for me!!!! PS: I got a reboot after the first line was done. I have windows 10. Above solutions are not worked for me. So I remove the `wrapt' by conda operation. conda remove wrapt And, it worked. pip install tensorflow it work for me. Thank you I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 It work for me. Thank you agreed! None of the solutions worked for me. I am always getting ERROR: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. and with conda remove wrapt I get RemoveError: 'setuptools' is a dependency of conda and cannot be removed from conda's operating environment. env: Mac OS 10.14.5 Above solutions are not worked for me. So I remove the `wrapt' by conda operation. conda remove wrapt And, it worked. pip install tensorflow Env: Windows 10 This worked for me. Tensorflow seems to have trouble being friendly with other packages. The following worked for me, starting from a clean conda environment: pip install --upgrade pip pip install six>=1.12.0 pip install httplib2==0.12.0 pip uninstall -y setuptools pip install setuptools>=41.0.0 pip install tensorflow==1.14.0 pip install tfx pip install pylint This sequence gets rid of wrapt as well as tensorboard warnings and google-apitools failures. This worked for me. I am using Jupyter notebook (OD: Ubuntu 16.0) all the commands running from the notebook: At the time of installing tensor flow I got following two errors. The errors and solutions are: ERROR 1: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. Solution: run following command !conda remove wrapt --yes ERROR 2: tensorboard 1.14.0 has requirement setuptools>=41.0.0 Solution: run following command to solve this error !pip uninstall -y setuptools !pip install setuptools>=41.0.0 Now install tensorflow: !pip install tensorflow dist-packages cannot be uninstalled or removed by pip, so you have to go manually delete the wrapt folder. Where is that wrapt folder?- It can be seen in the terminal from where it is picking the wrapt. in my case, the wrapt folder was in root/usr/lib/python3/ hope it works for you too. ps. (you can start by deleting all wrapt folders by just searching them) This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow It works for me. Thanks. conda update --all pip install tensorflow It works for me. thank you. It didn't work for me. Still getting that error. try conda update wrapt It didn't work for me. Still getting that error. try conda update wrapt It works for me. Thank you. It didn't work for me. Still getting that error. try conda update wrapt It worked for me. Thanks a lot. Above solutions are not worked for me. So I remove the `wrapt' by conda operation. conda remove wrapt And, it worked. pip install tensorflow Env: Windows 10 Thanks @NoSyu this work me This worked for me pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me too. pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me too,Thanks. Worked like charm Above solutions are not worked for me. So I remove the `wrapt' by conda operation. conda remove wrapt And, it worked. pip install tensorflow Env: Windows 10 it works for me, thanks very much! I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 Thanks..It worked for me. I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 works for me too!!!! I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 Worked for me too It didn't work for me. Still getting that error. Edit: Fixed it by installing tensorflow 1.14 first then updating it to version 2.0 This is is the only thing that worked for me. Thanks. Tried removing 'wrapt' and installing tensorflow v2.0 and then numpy threw a similar error. This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow it worked for me as well , thanks. This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow @rickykim93 you are awesome... I love you... This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow Hi, I had this problem in Raspberry Pi and your procedure worked for me. Thank you. Above solutions are not worked for me. So I remove the wrapt' by conda operation. conda remove wraptAnd, it worked.pip install tensorflow` Env: Windows 10 This worked for me. This also worked for me. Env: Mac OS 10.14.6 System information OS Platform and Distribution (e.g., Linux Ubuntu 19.04 on Sony notebook): TensorFlow version: python Python 3.6.3 |Anaconda, Inc.| (default, Oct 13 2017, 12:02:49) [GCC 7.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tensorflow as tf /home/ronald/anaconda3/lib/python3.6/site-packages/h5py/__init__.py:34: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`. from ._conv import register_converters as _register_converters >>> print(tf.__version__) 2.0.0-alpha0 Describe the problem I try to upgrade to beta1 with: install tensorflow==2.0.0-beta1 at the end I get: Installing collected packages: wrapt, tensorflow Found existing installation: wrapt 1.10.11 ERROR: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. How to fix it? if "pip install tensorflow" command won't work... try to "update conda --upgrade all" then "conda install tensorflow" this will definitely work. This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow Great fix. This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow Thanks alot, it works for me and able to installed TensorFlow and keras This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow it worked for me. thanks. This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow Works for me, thanks :) 它对我有用。谢谢 我通过以下方式修复它: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 它对我有用。谢谢 me too I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 This works for me, thanks! I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 worked...great..thanks This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow This works for my Windows 8.1 However, I was asked to consider upgrade the pip. And I did it first of all, using: python -m pip install --upgrade pip Thanks! pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me too in my mac,Thanks. This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow Thanks so much! This saved me. I fixed it by: conda update --all pip install --upgrade tensorflow==2.0.0-beta1 This worked for me conda update --all pip install tensorflow This worked for me Env: Mac OS 10.14.6 try conda from anaconda prompt or else it doesnt work This worked for me on a mac OS 10.14. Running in the terminal. pip install wrapt --upgrade --ignore-installed pip install tensorflow --upgrade I also fixed it by simply running conda update wrapt I think this avoids other unwanted upgrades by conda This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow @rickykim93 you are awesome... I love you... That worked for me too. Fantastic !!! This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow this worked for me. thanks a lot to the genius developers From all the solutions named here. This one is the only one that worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow This worked in Mac This worked for me: pip install wrapt --upgrade --ignore-installed pip install tensorflow I tried this and the tensorflow was successfully installed, but I had error when tired to use it. Than I tried pip install --upgrade tensorflow==2.0.0-beta1 and after this, it works well! That works for me conda uninstall wrapt pip install tensorflow Thank you man, this worked for me! Worked for Ubuntu 16.04 x86_64 When I perform the command "pip install tensorflow-gpu==1.15", I encounter the error "Found existing installation: wrapt 1.10.11 ERROR: Cannot uninstall 'wrapt'". my solution: perform "pip install wrapt --upgrade --ignore-installed", upgrade wrapt 1.10.11 to wrapt-1.12.0 perform "pip install tensorflow-gpu==1.15" if you cannot perform "pip install tensorflow-gpu==1.15" successfully. Maybe you firstly need to upgrade pip-18.1 to pip-20.0 using "python -m pip install --upgrade pip setuptools". This worked for me. I am using Jupyter notebook (OD: Ubuntu 16.0) all the commands running from the notebook: At the time of installing tensor flow I got following two errors. The errors and solutions are: ERROR 1: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. Solution: run following command !conda remove wrapt --yes ERROR 2: tensorboard 1.14.0 has requirement setuptools>=41.0.0 Solution: run following command to solve this error !pip uninstall -y setuptools !pip install setuptools>=41.0.0 Now install tensorflow: !pip install tensorflow It worked for me ,thank you! This worked for me. I am using Jupyter notebook (OD: Ubuntu 16.0) all the commands running from the notebook: At the time of installing tensor flow I got following two errors. The errors and solutions are: ERROR 1: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. Solution: run following command !conda remove wrapt --yes ERROR 2: tensorboard 1.14.0 has requirement setuptools>=41.0.0 Solution: run following command to solve this error !pip uninstall -y setuptools !pip install setuptools>=41.0.0 Now install tensorflow: !pip install tensorflow It worked for me ,thank you! Above solutions are not worked for me. So I remove the `wrapt' by conda operation. conda remove wrapt And, it worked. pip install tensorflow Env: Windows 10 It worked for me, thank you! This worked for me. I am using Jupyter notebook (OD: Ubuntu 16.0) all the commands running from the notebook: At the time of installing tensor flow I got following two errors. The errors and solutions are: ERROR 1: Cannot uninstall 'wrapt'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. Solution: run following command !conda remove wrapt --yes ERROR 2: tensorboard 1.14.0 has requirement setuptools>=41.0.0 Solution: run following command to solve this error !pip uninstall -y setuptools !pip install setuptools>=41.0.0 Now install tensorflow: !pip install tensorflow It worked for me , thank you It's worked for me as well. Really appreciated for providing such details. However, could you please provide more details regarding the problem? Why we are not able to uninstall "wrapt" without upgrading all the packages and libraries? @Elmit2015 Locking conversation to prevent a stream of "it works"
gharchive/issue
2019-06-27T03:51:28
2025-04-01T06:45:57.822542
{ "authors": [ "BhavishyaDiggiwal", "Boyombo1", "Devinmathew", "Elmit2015", "HaikunHuang", "Jingciii", "JuanYan", "KishaRenee", "MonsieurWave", "Neowy", "NoSyu", "PrasadSiri", "RealPolitiX", "RegaipKURT", "VRHaritha", "XiaoqianSun0104", "Yogeshnaik1190", "ajourney1", "atoyissan", "back-p", "bamwani", "bhabani-shankar", "chikum", "christophgm", "congee524", "deepeshguru", "eleprocha", "eric-0x72", "fabio11c", "fabriciofsantana", "gitvimalhub", "hi-im-ryanli", "houchengbin", "hridaydutta123", "hysunflower", "imoisharma", "javiervg", "jliu1235", "jmeile", "kentaroy47", "kurkutesa", "liangsun-ponyai", "lkfrota", "mihaimaruseac", "mikeyEcology", "moloudshahbazi", "nahidalam", "primrose3128", "qingxuantang", "rickykim93", "robertlugg", "scaler2017", "seyedmohammadmortaji", "ssusie", "sumiting", "tahwaru", "turinglife", "xiejunyu0531", "yangyang945", "zeineb12", "zung" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/30191", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
798591502
floating point exception in tf.nn.avg_pool3d and tf.nn.max_pool3dwhen ksize=0 System information Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 18.04 Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A TensorFlow installed from (source or binary): binary TensorFlow version (use command below):2.1.0 Python version:3.7.6 Bazel version (if compiling from source):N/A GCC/Compiler version (if compiling from source):N/A CUDA/cuDNN version:N/A GPU model and memory:N/A Describe the current behavior floating point exception in tf.nn.avg_pool3d and tf.nn.max_pool3d when ksize=0 Describe the expected behavior expect no crash Standalone code to reproduce the issue The code crashes in nightly version too. Check out the gist tf.nn.avg_pool3d(input=tf.ones((1,1,1,1,1)), strides=1, ksize=0, padding='VALID') tf.nn.max_pool3d(input=tf.ones((1,1,1,1,1)), strides=1, ksize=0, padding='VALID') Output: Floating point exception (core dumped) @ravikyram I have created a pull request to solve this issue , which will help raise FloatingPoitnError when ksize = 0 I have tried in colab with TF version 2.4, nightly version(2.5.0-dev20210201) and was able to reproduce the issue.Please, find the gist here. Thanks!
gharchive/issue
2021-02-01T18:12:40
2025-04-01T06:45:57.829912
{ "authors": [ "DNXie", "around-star", "ravikyram" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/46834", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
181759336
Android Demo build error The error I saw is as follows, don't quit know how this permission error occurs ERROR: /home/toxido/.cache/bazel/_bazel_root/e7444205eb7a78f2fb0f6da55e24d78b/external/androidsdk/BUILD:5:1: Executing genrule @androidsdk//:aapt_runner failed: linux-sandbox failed: error executing command /home/toxido/.cache/bazel/_bazel_root/e7444205eb7a78f2fb0f6da55e24d78b/execroot/tensorflow/_bin/linux-sandbox ... (remaining 5 argument(s) skipped). src/main/tools/linux-sandbox-pid1.cc:233: "mount(/, /home/toxido/.cache/bazel/_bazel_root/e7444205eb7a78f2fb0f6da55e24d78b/bazel-sandbox/28d014be-3643-4295-b588-2ed7789511a4-1/tmp, NULL, MS_BIND | MS_REC, NULL)": Permission denied Target //tensorflow/examples/android:tensorflow_demo failed to build Use --verbose_failures to see the command lines of failed build steps. @ToxidoLiu Please fill out the following information, which was the default template when you clicked on the green "New issue" button. Also include the exact command-line that you ran, that causes the error. What related GitHub issues or StackOverflow threads have you found by searching the web for your problem? Environment info Operating System: Installed version of CUDA and cuDNN: (please attach the output of ls -l /path/to/cuda/lib/libcud*): If installed from binary pip package, provide: A link to the pip package you installed: The output from python -c "import tensorflow; print(tensorflow.__version__)". If installed from source, provide The commit hash (git rev-parse HEAD) The output of bazel version If possible, provide a minimal reproducible example (We usually don't have time to read hundreds of lines of your code) What other attempted solutions have you tried? Logs or other output that would be helpful (If logs are large, please upload as attachment or provide link). @tatatodd OS & Environment : Ubuntu 16.04 LTS , Android Studio v2.2 tensorflow version : 0.8.0 bazel version: 0.3.2 Android SDK version: 24.0.3 NDK version: 13.0.3315539 The above error occured when I tried to do '$ bazel build //tensorflow/examples/android:tensorflow_demo' as instructed here: https://github.com/tensorflow/tensorflow/tree/master/tensorflow/examples/android Thanks! @ToxidoLiu Please try the latest version of TensorFlow 0.11, since 0.8 is quite old at this point: https://www.tensorflow.org/versions/r0.11/get_started/os_setup.html#download-and-setup @toxidoliu Have you checked that relevant binaries all have the execute bit set? If you're still having issues after you upgrade TF can you paste the error log running with --verbose_failures? Closing automatically due to lack of recent activity. Please reopen when further information becomes available. Thank you.
gharchive/issue
2016-10-07T20:21:08
2025-04-01T06:45:57.837675
{ "authors": [ "ToxidoLiu", "andrewharp", "aselle", "tatatodd" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/4832", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
876408178
Dynamic Input support for TFLite models Hello Tensors I have a TF model with (None,None,100) input dims and its converted into TFLite too. But when I do inference, it gives error- RuntimeError: tensorflow/lite/kernels/reshape.cc:58 stretch_dim != -1 (0 != -1)Node number 72 (RESHAPE) failed to prepare. System information TensorFlow version (you are using): 2.4.1. Describe the feature and the current behavior/state. Currently, 0th index can be dynamic(None or -1). Will this change the current api? How? Other indexs of model input can be dynamic(None or -1) Who will benefit with this feature? Generally, Vision models have 0th index dynamic but NLP models have other dimensions as dynamic. @ichuang @neso613 Please share simple stand alone code such that we can replicate the issue faced or if possible share a colab gist. FYI, both TF and TFLite supports multiple dimensions but the Reshape operator does not only allow two dynamic dimensions in the shape argument. @neso613 Please share simple stand alone code such that we can replicate the issue faced or if possible share a colab gist. FYI, both TF and TFLite supports multiple dimensions but the Reshape operator does not only allow two dynamic dimensions in the shape argument. Thats the Question. TFLite supports dynamic input shape. [{'name': 'input_1', 'index': 0, 'shape': array([ 1, 1, 40], dtype=int32), 'shape_signature': array([-1, -1, 40], dtype=int32), 'dtype': <class 'numpy.float32'>, 'quantization': (0.0, 0), 'quantization_parameters': {'scales': array([], dtype=float32), 'zero_points': array([], dtype=int32), 'quantized_dimension': 0}, 'sparsity_parameters': {}}] When I do inference, my input can be of any shape, so I have to resize using interpreter.resize_tensor_input(input_details[0]['index'],data_test.shape) That is an intended behavior since the input tensors should be fixed through the resize_tensor_input method when the given model has dynamic inpiuts. You can resize input tensors again when the model should handle the different size in the following tasks. That is an intended behavior since the input tensors should be fixed through the resize_tensor_input method when the given model has dynamic inpiuts. You can resize input tensors again when the model should handle a different size in the following tasks. This is the code, I am using interpreter.resize_tensor_input(input_details[0]['index'],data_test.shape) interpreter.allocate_tensors() interpreter.set_tensor(input_details[0]['index'], data_test) interpreter.invoke() output_data = interpreter.get_tensor(output_details[0]['index']) @neso613 Please share simple stand alone code such that we can replicate the issue faced or if possible share a colab gist. It is really impossible to reproduce your problem at our end and please consider the latest TF version from the conversion stage to the inference stage. I have gotten a very similar error when converting models with TFLite and running with TensorFlow 2.4. TensorFlow versions 2.3 and 2.5 pre-release work for me, I'd recommend trying these versions. It is running now, but with Flex dependecny. Any solution for this flex issue? Please file a feature request for the Select TF ops by adding a new issue. Thanks.
gharchive/issue
2021-05-05T12:40:56
2025-04-01T06:45:57.846196
{ "authors": [ "abattery", "gcervantes8", "neso613" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/48915", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2707004519
[-Viral@Link-]— Muskan Chandio viral original Video Link Muskan Chandio xxl Video Viral Updated Link On Social Media X Trending Now.L𝚎aked Video Muskan Chandio X Original Video Viral Video L𝚎aked on X Twitter Telegram. 𝖫𝖾𝖺𝚔𝖾𝖽 𝗏𝗂𝗋𝖺𝗅 𝗅𝗂𝗇𝗄 2024 𝖫𝖾𝖺𝚔𝖾𝖽 𝖵𝗂𝖽𝖾𝗈.Muskan Chandio Full Sex Video Original Video Link Muskan Chandio Viral On Social Media X Trending Now.Muskan Chandio Sex Video Link Short Clip Muskan Chandio Video Viral On Social Media X TikTok Trending 20 seconds ago L𝚎aked Video Muskan Chandio Original Video Viral Video L𝚎aked on X Twitter Telegram .. .. 🌐 𝖢𝖫𝖨𝖢𝖪 𝖧𝖤𝖱𝖤 🟢==►► 𝖶𝖠𝖳𝖢𝖧 𝖭𝖮𝖶 .. .. 🔴 𝖢𝖫𝖨𝖢𝖪 𝖧𝖤𝖱𝖤 🌐==►► 𝖣𝗈𝗐𝗇𝗅𝗈𝖺𝖽 𝖭𝗈𝗐 .. .. .. .. [-wATCH-]— Muskan Chandio Video Original Video Link Muskan Chandio Video Viral On Social Media X Trending Now [-wATCH-]— Muskan Chandio ʟᴇᴀᴋᴇᴅ Video ᴠɪʀᴀʟ On Social Media ˣ ᵀʷⁱᵗᵗᵉʳ [-wATCH-]— Muskan Chandio ʟᴇᴀᴋᴇᴅ Video ᴠɪʀᴀʟ On Social Media ˣ ᵀʷⁱᵗᵗᵉʳ [-wATCH-]— Muskan Chandio Video Original Video Link Muskan Chandio Video Viral On Social Media X Trending Now Muskan Chandio Original Video video took the internet by storm and amazed viewers on various social media platforms. Muskan Chandio, a young and talented digital creator, recently became famous thanks to this interesting video. L𝚎aked Video Muskan Chandio Original Video Viral Video L𝚎aked on X Twitter Muskan Chandio Original Video video oficial twitter L𝚎aked Video Muskan Chandio Original Video Viral Video L𝚎aked on X Twitter... . . . . . 20 seconds ago L𝚎aked Video Muskan Chandio Original Video Viral Video L𝚎aked on X Twitter Telegram .. .. 🌐 𝖢𝖫𝖨𝖢𝖪 𝖧𝖤𝖱𝖤 🟢==►► 𝖶𝖠𝖳𝖢𝖧 𝖭𝖮𝖶 .. .. 🔴 𝖢𝖫𝖨𝖢𝖪 𝖧𝖤𝖱𝖤 🌐==►► 𝖣𝗈𝗐𝗇𝗅𝗈𝖺𝖽 𝖭𝗈𝗐 .. .. .. .. [-wATCH-]— Muskan Chandio Video Original Video Link Muskan Chandio Video Viral On Social Media X Trending Now [-wATCH-]— Muskan Chandio ʟᴇᴀᴋᴇᴅ Video ᴠɪʀᴀʟ On Social Media ˣ ᵀʷⁱᵗᵗᵉʳ [-wATCH-]— Muskan Chandio ʟᴇᴀᴋᴇᴅ Video ᴠɪʀᴀʟ On Social Media ˣ ᵀʷⁱᵗᵗᵉʳ [-wATCH-]— Muskan Chandio Video Original Video Link Muskan Chandio Video Viral On Social Media X Trending Now Muskan Chandio Original Video video took the internet by storm and amazed viewers on various social media platforms. Muskan Chandio, a young and talented digital creator, recently became famous thanks to this interesting video. L𝚎aked Video Muskan Chandio Original Video Viral Video L𝚎aked on X Twitter Muskan Chandio Original Video video oficial twitter L𝚎aked Video Muskan Chandio Original Video Viral Video L𝚎aked on X Twitter.. BOL EntertainmentAsh Muskan Chandio Video Tutorial Original Video video took the internet by storm and amazed viewers on various social media platforms. . . . . . . . . . . . . . .. . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . dsa 1st . . . . . . . . . . . . . . . . . . . . . . 2nd . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . cvcv
gharchive/issue
2024-11-30T08:42:51
2025-04-01T06:45:57.894507
{ "authors": [ "amiraufaa", "jayasanka009", "lailihasan6", "mismebasu" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/81411", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2709238437
Destiny y Nick Fuentes Filtrado Destiny y Nick Fuentes Viral En Redes Sociales X TikTok Tendencia Hace 44 segundos Video visto Video original de TIKTOKER PAKISTANÍ Video viral visto en X Twitter Telegram " " "🌐 HAZ CLIC AQUÍ 🟢==►► MIRA AHORA " " " 🔴 HAZ CLIC AQUÍ 🌐==►► Descargar ahora " " " 𝖫𝖾𝖺𝚔𝖾𝖽 𝗏𝗂𝗋𝖺𝗅 𝗅𝗂𝗇𝗄 2024 𝖫𝖾𝖺𝚔𝖾𝖽 𝖵𝗂𝖽𝖾𝗈 𝖵𝗂𝗋𝖺𝗅 𝖬𝖾𝖽𝖾𝖼𝗂𝖺𝗅 𝖬𝖾𝖽𝗂𝖺 2024.𝖫𝖾𝖺𝚔𝖾𝖽 𝗏𝗂𝗋𝖺𝗅 𝗅𝗂𝗇𝗄 2024 𝖫𝖾𝖺𝚔𝖾𝖽 𝖵𝗂𝖽𝖾𝗈 𝖵𝗂𝗋𝖺𝗅 𝖮𝗇 𝖲𝗈𝖼𝗂𝖺𝗅 Año 2024. El calendario lunar de hoy es de 2021. 𝖿𝖺𝗅𝗅𝗈𝗎𝗍𝗈 𝗌𝖾𝗋𝗏𝖾𝗌 𝖺𝗌 𝖺 𝗋𝖾𝗆𝗂𝗇𝖽𝖾𝗋 𝗈𝖿 𝗍𝗁𝖾 𝖿𝗋𝖺𝗀𝗂𝗅𝗂𝗍𝗒 𝗈𝖿 𝗋𝖾𝗉𝗎𝗍𝖺𝗍𝗂𝗈𝗇 𝗂𝗇 𝗍𝗁𝖾 𝖽𝗂𝗀𝗂𝗍𝖺𝗅 𝖾𝗋𝖺. 𝖠𝗌 𝗍𝗁𝖾 𝗅𝗂𝗇𝖾𝗌 𝖻𝖾𝗍𝗐𝖾𝖾𝗇 𝗉𝗋𝗂𝗏𝖺𝗍𝖾 𝖺𝗇𝖽𝖽 𝗉𝗎𝖻𝗅𝗂𝖼 𝗅𝗂𝖿𝖾 𝖼𝗈𝗇𝗍𝗂𝗇𝗎𝖾 𝗍𝗈 𝖻𝗅𝗎𝗋, 𝖼𝖾𝗅𝖾𝖻𝗋𝗂𝗍𝗂𝖾𝗌 𝗅𝗂𝗄𝖾 𝖯𝗋𝗂𝗌𝗈𝗇 𝖮𝖿𝖿𝗂𝖼𝖾𝗋𝖿𝗂𝗇𝖽𝖾 𝗍𝗁𝖾𝗆𝗌𝖾𝗅𝗏𝖾𝗌 𝖺𝗍 𝗍𝗁𝖾 𝗆𝖾𝗋𝖼𝗒 𝗈𝖿 𝗗𝗘 𝗖𝗜𝗡𝗧𝗘𝗥𝗖𝗜Ó𝗡 𝗗𝗘 𝗗𝗘 𝗗𝗘𝗡𝗧𝗘𝗥𝗖𝗜Ó𝗡 𝗗𝗘 𝗗𝗘 𝗔𝗡𝗧𝗘𝗥𝗖𝗜Ó𝗡 ... 𝖺𝗇𝖽 𝗃𝗎𝖽𝗀𝗆𝖾𝗇𝗍𝗈. 𝖨𝗇 𝗍𝗁𝖾 𝖾𝗏𝖾𝗋 𝖾𝗏𝗈𝗅𝗏𝗂𝗇𝗍𝗈 𝗅𝖺𝗇𝖽𝗈𝗌𝖼𝖺𝗉𝖾 𝗈𝖿 𝖼𝖾𝗅𝖾𝖻𝗋𝗂𝗍𝗒 𝖼𝗎𝗅𝗍𝗎𝗋𝖾, 𝗍𝗁𝖾 𝖨𝗌𝗁𝗈𝗐𝗌𝗉𝖾𝖾𝖽𝗌𝖼𝖺𝗇𝖽𝖺𝗅 𝗎𝗇𝖽𝖾𝗋𝗌𝖼𝗈𝗋𝖾𝗌 𝗍𝗁𝖾 𝗋𝖾𝗅𝖾𝗇𝗍𝗅𝖾𝗌𝗌 𝗉𝗎𝗋𝗌𝗎𝗂𝗍𝗈 𝗌𝖾𝗇𝗌𝖺𝗍𝗂𝗈𝗇𝖺𝗅𝗂𝗌𝗆𝗈, 𝖺 𝗉𝗎𝗋𝗌𝗎𝗂𝗍 𝗍𝗁𝖺𝗍 𝗈𝖿𝗍𝖾𝗇 𝖼𝗈𝗆𝖾𝗌 𝖺𝗍 𝗍𝗁𝖾 𝖾𝗑𝗉𝖾𝗇𝗌𝖾 𝗈𝖿 𝗍𝗋𝗎𝗍𝗁 𝖺𝗇𝖽𝖺 𝖽𝗂𝗀𝗇𝗂𝗍𝗒. 𝖠𝗌 𝗐𝖾 𝗇𝖺𝗏𝗂𝗀𝖺𝗍𝖾 𝗍𝗁𝖾 𝖼𝗈𝗆𝗉𝗅𝖾𝗑𝗂𝗍𝗂𝖾𝗌 𝗈𝖿 𝗍𝗁𝖾 𝖽𝗂𝗀𝗂𝗍𝖺𝗅 𝖺𝗀𝖾, 𝗍𝗁𝖾 𝗅𝗂𝗇𝖾 El programa de fidelización de veteranos de guerra de 2018 está en marcha. 𝖠𝗌 𝗔𝗻 𝗹𝗶𝗻𝗲𝗻 𝗲𝗻 𝗹𝗮 𝗽𝗮𝗿𝗮 ...�𝗼𝗺𝗽𝗮𝗿𝗮 El programa de fidelización de veteranos de la Segunda Guerra Mundial está en marcha. Yo estaba en el hospital y estaba muy ocupado con mis tareas, así que, 𝗌𝗀𝗶𝗻 𝗲𝗻 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗽𝗮𝗿𝗮 𝗽𝗮𝗿𝗮 𝗹𝗮 𝗗𝗘 𝗖𝗜𝗡𝗜𝗢 𝗗𝗘 𝗗𝗘 𝗖𝗜𝗢𝗡𝗖𝗜Ó𝗡 𝗗𝗘 𝗔𝗡𝗧𝗘𝗥𝗖𝗜Ó𝗡 𝗗𝗘 𝗔𝗡𝗧𝗘𝗥𝗧𝗘𝗦 𝗗𝗘 𝗉𝗋𝗂𝗏𝖺𝗼𝗒 𝗗𝗘 𝗔𝗡𝗧𝗘𝗥𝗧𝗘𝗦. cvcv
gharchive/issue
2024-12-01T14:23:15
2025-04-01T06:45:57.898223
{ "authors": [ "ba05251670", "semolisk" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/issues/81811", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
347565992
fix var type issue which breaks crf_decode CRF decode can fail when default type of "0" (as viewed by math_ops.maximum) does not match the type of sequence_length. This change is parallel in motivation and solution to the fix in https://github.com/tensorflow/tensorflow/commit/c0f1080188c5c6955cfa3b3c086ac262b1e5ec02, for crf_log_norm()=>_multi_seq_fn(). Just modified one of the a couple of the CRF tests to presumably(?) provide better coverage of different var types for sequence_length settings.
gharchive/pull-request
2018-08-03T23:58:47
2025-04-01T06:45:57.901462
{ "authors": [ "cbockman" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/pull/21371", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2296978713
Fix ConcreteFunction when called with structured outputs. ConcreteFunction properly flattens structured inputs (e.g. composite tensors) when called, but it does not pack/un-flatten the outputs. When ConcreteFunction outputs are structured, they must be packed back into their structured form to match the function signature. Hi @james-choncholas Can you please sign the CLA? Thank you! Force-pushed to update the commits email address to the one associated with my Google contributor CLA. Hi @rohan100jain Any update on this PR? Please. Thank you! Hi @rohan100jain Any update on this PR? Please. Thank you! Hi @rohan100jain Any update on this PR? Please. Thank you! Hi @rohan100jain Any update on this PR? Please. Thank you! Hi @rohan100jain, Any update on this PR? Please. Thank you! Hi @rohan100jain, Any update on this PR? Please. Thank you! Hi @rohan100jain, Any update on this PR? Please. Thank you! Hi @rohan100jain, Any update on this PR? Please. Thank you!
gharchive/pull-request
2024-05-15T06:29:19
2025-04-01T06:45:57.904903
{ "authors": [ "gbaned", "james-choncholas", "keerthanakadiri" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/pull/67612", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
222313675
Explicitly delete Estimator before copying output_dir Hopefully this fixes #9185 by releasing some stray file handle. I triggered http://ci.tensorflow.org/job/tensorflow-pr-win-bazel/3/console I hope the setup is the same as tf-master-win-bzl. @meteorcloudy FYI. @martinwicke I tested on my local machine as well, it doesn't seem to fix the problem. But I found shutil.copytree(model_dir1, model_dir2) works, can we use this instead of os.rename? The problem with that would be that this does not delete the old tree. This test is to ensure that no absolute paths creep in anywhere, so this would defang the test. I cannot see what hangs on to a file handle. Perhaps the easiest approach would be to install the current nightly on one of our Windows test VMs, run a modified version of estimator_test.py that blocks before the rename, and apply https://technet.microsoft.com/en-us/sysinternals/processexplorer.aspx to the Python interpreter to see what handles are open? @mrry It shows this file is in use by python c:/tmp/tmp3y2vhvzi/model_dir1/events.out.tfevents.1492671112.TENSORFLOW-JENK Thanks for tracking that down, Yun! Then I think we have a smoking gun: the tf.summary.FileWriter is holding open a file handle after est1.train() returns. It looks like this happens deep in the MonitoredTrainingSession... and there seems to be a cache(!!) involved: https://github.com/tensorflow/tensorflow/blob/master/tensorflow/python/summary/writer/writer_cache.py#L27 As a hack, you might try invoking FileWriterCache.clear() before the rename, but I'm not sure that that's sufficient because deleting a FileWriter doesn't appear to close() it, so you might need to add a writer.close() call somewhere (or maybe add it to the destructor?). @mrry Thanks for the hint, Derek! I tested adding writer_cache.FileWriterCache._cache[model_dir1].close() before rename also solves the problem. Does this sounds good to you? That'll fix the test, but I'm wondering whether we can get put that line (or some equivalent) somewhere were it will be called automatically when the Estimator owning the model_dir is deleted? I'm even wondering whether we should even keep the cache live when any of train(), eval(), predict() exit. They're all meant to be long running and ideally shouldn't leave open file handles hanging around. @martinwicke Yes, I totally agree! But I am lack of knowledge how these function works. Can someone with more experience look into this? Jenkins, test this please. Triggered another Windows build: http://ci.tensorflow.org/job/tensorflow-pr-win-bazel/4/ I edited FileWriterCache to close() all caches it has, and call that in the test. Cleaner than closing the cache directly. I'll merge to fix build.
gharchive/pull-request
2017-04-18T06:19:07
2025-04-01T06:45:57.913619
{ "authors": [ "martinwicke", "meteorcloudy", "mrry" ], "repo": "tensorflow/tensorflow", "url": "https://github.com/tensorflow/tensorflow/pull/9281", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
309426689
Starter Project links broken Starter project links at homepage are broken https://deeplearnjs.org/index.html#demos https://github.com/PAIR-code/deeplearnjs/tree/master/starter/es6/ https://github.com/PAIR-code/deeplearnjs/tree/master/starter/typescript/ Apologies, we're in the middle of moving some code to other repositories. We'll send an update on Friday with where all the code now lives! Hi @kriptontr We are moving demos/starter projects to a new repo with the launch of https://js.tensorflow.org/ See https://github.com/tensorflow/tfjs-examples for the things start off from. With this we are removing starter projects from the main repository. We are keeping up the old site for a while to make the transition easier but it is essentially deprecated.
gharchive/issue
2018-03-28T15:51:30
2025-04-01T06:45:57.917020
{ "authors": [ "kriptontr", "nsthorat", "tafsiri" ], "repo": "tensorflow/tfjs-core", "url": "https://github.com/tensorflow/tfjs-core/issues/906", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2123778609
Consumer core gives different NoC dst for matmul To reproduce: go to yugao/wh_block_matmul_hang_consumer run pytest tests/tt_eager/python_api_testing/unit_testing/test_bert_ops.py In tt_metal/impl/dispatch/kernels/command_queue_consumer.hpp, I added DPRINT info for command_ptr[3], noc_x_start, noc_y_start, noc_x_end, noc_y_end. The command_ptr[3] should contain the NoC dst. The branch is based on the yugao/wh_block_matmul_hang, so it will han after running the test for around 100 iterations. Machine IP: 172.27.44.23, harvested X2 @yugaoTT is this a blocker? @davorchap no I was debugging the matmul_hang, and saw the consumer is mcasting go signal to a NoC dst of (1,3),(9,9) for matmul, while for interleaved_to_sharded the consumer is mcasting to (1,2),(9,9), which is a grid of 8x7. I looked into this, and for matmul, given it is broken into two kernel groups, it is correct that Yu would see two GO signal mcasts. One potential bug we did catch though is that there is nothing stalling the dispatcher on sending the GO signals in the case where we have multiple kernel groups. That means it is possible for semaphores to not arrive to one group before another group gets its GO signal. Apparently this is not done for this matmul kernel, but it seems like a general problem. Ideally, we'd want to stall sending the GO signal before binaries/semaphores have arrived in cases where we have two kernel groups. @pgkeller @davorchap FYI. Closing this as a non-issue, will spin up another issue for the race that Yu and I caught.
gharchive/issue
2024-02-07T20:02:24
2025-04-01T06:45:58.140811
{ "authors": [ "DrJessop", "davorchap", "yugaoTT" ], "repo": "tenstorrent-metal/tt-metal", "url": "https://github.com/tenstorrent-metal/tt-metal/issues/5189", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2226546240
FD2: PrefetchQ possible corruption Running Finish stress test we see: prefetch q is all zeros host writes an entry device sees the entry but also sees the prior entry get re-written Things to try: add sfence after each write to fetchQ StrictOrdering on TLB 16B fetchQ entry make L1 uncachable @davorchap @imatosevic @tt-aho @tooniz @abhullar-tt @ubcheema @aliuTT Additional note is this was seen on a single GS machine. Typically this would occur under ~2 million loops. Tried on a WH on a T3000 machine and didn't see an issue after ~5 million iterations before stopping it. Can retry on WH as well to confirm. Additional note is this was seen on a single GS machine. Typically this would occur under ~2 million loops. Tried on a WH on a T3000 machine and didn't see an issue after ~5 million iterations before stopping it. Can retry on WH as well to confirm. How long does 2 million loops take? Yes, maybe run it over night.
gharchive/issue
2024-04-04T21:25:52
2025-04-01T06:45:58.144508
{ "authors": [ "pgkeller", "tt-aho" ], "repo": "tenstorrent-metal/tt-metal", "url": "https://github.com/tenstorrent-metal/tt-metal/issues/7136", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2070828301
Adding interface to assign dispatch kernels to dispatch functionality and adding kernel to service remote command queue Host changes create dispatch core assigner interface to track which dispatch cores are assigned to which dispatch functionalities because with multi-chip fast dispatch we have more dispatch functionality than just cq producer and cq consumer (defined in dispatch_core_manager.hpp) update SendDispatchKernelsToDevice to program remote cq interface cores as well as mmio dispatch cores when initializing mmio device update CQ interface apis to account for hugepage channel offsets (each device has a hugepage in different channel) Device changes added remote_issue_queue_interface kernel to pull commands from remote command queue. In this PR that kernel doesn't do anything but @aliuTT is working on integrating it to the ethernet core Passing post commit test suite
gharchive/pull-request
2024-01-08T16:34:33
2025-04-01T06:45:58.147478
{ "authors": [ "abhullar-tt" ], "repo": "tenstorrent-metal/tt-metal", "url": "https://github.com/tenstorrent-metal/tt-metal/pull/4615", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2103706984
#0: Fix dprint server lock not being released before throw Manually verified that the test fails on my BM before this fix, and that is passes after this fix lgtm @aliuTT has a PR that includes this, will close this one without merging
gharchive/pull-request
2024-01-27T18:14:36
2025-04-01T06:45:58.148642
{ "authors": [ "aliuTT", "tt-dma" ], "repo": "tenstorrent-metal/tt-metal", "url": "https://github.com/tenstorrent-metal/tt-metal/pull/4992", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2267658366
#7919: fix barriers in s2i writers and add read_one_packet for row-ma… …jor wormhole case I see. I dunno about all cases but it did help cases in RN50. (From 100k cycles to 80k cycles). Ill put the barriers back and investigate this more throughly in a separate PR. On Mon, Apr 29, 2024 at 09:08 Austin Ho @.***> wrote: @tarafdarTT https://github.com/tarafdarTT do you have profiling data that moving the barriers has improved perf in all cases? When this op was originally brought up, then having the barrier outside the inner loop was sometimes suboptimal due to noc congestion — Reply to this email directly, view it on GitHub https://github.com/tenstorrent/tt-metal/pull/7920#issuecomment-2082695744, or unsubscribe https://github.com/notifications/unsubscribe-auth/BAK3IAYLCGHUTRPO2PCA6W3Y7ZA3XAVCNFSM6AAAAABG5C4CBGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDAOBSGY4TKNZUGQ . You are receiving this because you were mentioned.Message ID: @.***> I used the default vim formatter. On Mon, Apr 29, 2024 at 09:15 Austin Ho @.***> wrote: @.**** commented on this pull request. In tt_eager/tt_dnn/op_library/sharded/kernels/dataflow/reader_unary_stick_layout_sharded_blocks_interleaved_start_id.cpp https://github.com/tenstorrent/tt-metal/pull/7920#discussion_r1583070950 : @@ -7,57 +7,83 @@ void kernel_main() { const uint32_t src_addr = get_arg_val<uint32_t>(0); What kind of code formatter did you use on these kernel files? We normally use clang-format if we format code and this doesn't seem to match. — Reply to this email directly, view it on GitHub https://github.com/tenstorrent/tt-metal/pull/7920#pullrequestreview-2028483098, or unsubscribe https://github.com/notifications/unsubscribe-auth/BAK3IA2FWSTUATLMQGNEF7LY7ZBWXAVCNFSM6AAAAABG5C4CBGVHI2DSMVQWIX3LMV43YUDVNRWFEZLROVSXG5CSMV3GSZLXHMZDAMRYGQ4DGMBZHA . You are receiving this because you were mentioned.Message ID: @.***> I used the default vim formatter. … On Mon, Apr 29, 2024 at 09:15 Austin Ho @.> wrote: @.* commented on this pull request. ------------------------------ In tt_eager/tt_dnn/op_library/sharded/kernels/dataflow/reader_unary_stick_layout_sharded_blocks_interleaved_start_id.cpp <#7920 (comment)> : > @@ -7,57 +7,83 @@ void kernel_main() { - const uint32_t src_addr = get_arg_val<uint32_t>(0); What kind of code formatter did you use on these kernel files? We normally use clang-format if we format code and this doesn't seem to match. — Reply to this email directly, view it on GitHub <#7920 (review)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/BAK3IA2FWSTUATLMQGNEF7LY7ZBWXAVCNFSM6AAAAABG5C4CBGVHI2DSMVQWIX3LMV43YUDVNRWFEZLROVSXG5CSMV3GSZLXHMZDAMRYGQ4DGMBZHA . You are receiving this because you were mentioned.Message ID: @.***> Could you either revert the formatting or use clang-format? The vim formatting does not match our style guide
gharchive/pull-request
2024-04-28T16:08:46
2025-04-01T06:45:58.161505
{ "authors": [ "tarafdarTT", "tt-aho" ], "repo": "tenstorrent/tt-metal", "url": "https://github.com/tenstorrent/tt-metal/pull/7920", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2275177048
#7956: Fix addc like ttnn ops doc Fix issue #7956 Add supported dtypes and layouts and skipped that are unsupported Reflect the same in documentation Sweep Test Results: addcdiv.csv addcmul.csv All post commit test : https://github.com/tenstorrent/tt-metal/actions/runs/8922478405 Fix merged to main
gharchive/pull-request
2024-05-02T10:35:26
2025-04-01T06:45:58.164109
{ "authors": [ "Aswinmcw" ], "repo": "tenstorrent/tt-metal", "url": "https://github.com/tenstorrent/tt-metal/pull/8045", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2278084992
#0: umd bump BH support in UMD is coming soon, so we'll need to bump sometime in the near future This is a test to see if anything broke ???
gharchive/pull-request
2024-05-03T17:02:59
2025-04-01T06:45:58.165072
{ "authors": [ "TT-billteng", "tt-rkim" ], "repo": "tenstorrent/tt-metal", "url": "https://github.com/tenstorrent/tt-metal/pull/8104", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2352217688
#8482: Trace on remote chips Ticket https://github.com/tenstorrent/tt-metal/issues/8482 Problem description Trace is not working on remote chips. What's changed Changes from @pgkeller: Fix host completion write ptr in test_prefetcher Was clobbering cmds. This enables more prefetcher test configs w/ exec_buf Rework split prefetch exec_buf sync mechanism prefetch_h now stalls itself at the start of exec_buf by taking away its downstream page credits. After sending the exec_buf, it unstalls the prefetch_q so that it can continue to do prefetch work. dispatch restores prefetch_h credits at exec_buf_end. Increase test_prefetcher exec_buf test coverage cq_prefetch changes: Modify detection of exec_buf command. Prefetcher no longer looks at the cmd_id, since exec_buf command is not the first command in the Fetch Q. Instead rely on stall flag. prefetch_h should not offset exec_buf command by preamble size, since it belongs to the same entry as a previously offset command (due to barrier_and_stall) Account for wrap in barrier and stall Add T3K trace tests. Checklist [x] Post commit CI passes [ ] Model regression CI testing passes (if applicable) [x] New/Existing tests provide coverage for changes can you run the t3k frequent tests pipeline to test your changes? can you run the t3k frequent tests pipeline on your branch to test your changes? will approve if the specific test suite in t3k frequent CI pipeline tests pass Done, trace stress tests were passing
gharchive/pull-request
2024-06-13T23:36:09
2025-04-01T06:45:58.171156
{ "authors": [ "tt-asaigal", "ttmchiou" ], "repo": "tenstorrent/tt-metal", "url": "https://github.com/tenstorrent/tt-metal/pull/9433", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1253633676
Add Internal Load Balancer This change wraps the VAST server task in a load balancer into a service and adds a network load balancer in front. This allows us to: expose the server on a pre-defined local IP, removing the need to dynamically resolve it through the Fargate API automatically restart the VAST server when it crashes It requires a few extra resources to be initiated: in terms of code complexity, this is offset-ed by the fact we can remove all the logic for the dynamic IP resolution in terms of cost, the extra network load balancer adds a fix cost of ~$0.03 / hour to the stack :memo: Checklist [ ] All user-facing changes have changelog entries. [ ] The changes are reflected on docs.tenzir.com/vast, if necessary. [x] The PR description contains instructions for the reviewer, if necessary. :dart: Review Instructions Add a service and a load balancer resource in the fargate module The task now being wrapped in a service, we use desiredCount to start / stop the server instead of start-task / stop-task API calls (for stopping, we keep stop-task to speed up the process) Also did some minor changes in the tagging of the network resources because they were detected as perpetual changes by Terraform simplified the integration tests to avoid having to do long cleanups of the EFS drive Needs to be rebased once https://github.com/tenzir/vast/pull/2315 is merged before I changed the resource naming strategy a bit because they were sometimes too long, integration tests should now pass: https://github.com/rdettai/vast/runs/6774453785 Note that this PR should go with some minor updates of the setup README, but we have batched them into https://github.com/tenzir/vast/pull/2322 instead. I think this is ready to merge @dispanser !
gharchive/pull-request
2022-05-31T10:03:29
2025-04-01T06:45:58.180164
{ "authors": [ "rdettai" ], "repo": "tenzir/vast", "url": "https://github.com/tenzir/vast/pull/2319", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1414466839
Add a --web-root option to the web plugin :memo: Reviewer Checklist Review this pull request by ensuring the following items: [ ] All user-facing changes have changelog entries [ ] User-facing changes are reflected on vast.io Also, that CORS error should go away with this patch diff --git a/src/routes/status/+page.svelte b/src/routes/status/+page.svelte index 9bbed5b..07c081d 100644 --- a/src/routes/status/+page.svelte +++ b/src/routes/status/+page.svelte @@ -14,7 +14,7 @@ const getStatus = async () => { try { - const API_BASE = 'http://localhost:42001/api/v0'; + const API_BASE = 'http://127.0.0.1:42001/api/v0'; const url = `${API_BASE}/status?verbosity=detailed`; const response = await fetch(url); Historically, the "web root" is the place where a web server places its file, so --web-root to me feels like a well-known thing. I have no strong opinions on calling it differently though. My mistake. I thought web/docs, web/blog and plugins/web/ui are at the same level, and are also served from the CLI - so I wanted to disambiguate. But since that is not the case, --web-root is perfectly fine. I changed it to just --root for now, since that's what nginx also uses in their configuration and there doesn't seem to be consensus between "document root" and "web root" between different web servers. I "merged" this manually by rebasing and pushing onto the base branch, to keep the commit graph a bit simpler by avoiding multiple merge bases.
gharchive/pull-request
2022-10-19T07:39:57
2025-04-01T06:45:58.184259
{ "authors": [ "dit7ya", "lava", "mavam" ], "repo": "tenzir/vast", "url": "https://github.com/tenzir/vast/pull/2641", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1503688569
Fix UB in order(...) for floating point numbers The coder unit tests using the order(...) function on floating point numbers started failing for release builds on Debian after the CAF 0.18 bump, but the unit test suite wasn't touched since 2016, and the code itself not since 2018, so it was pretty like to be undefined behavior that we ran into. That turned out be rather easy to spot looking at the function. But do you need to launder the pointer? Laundering suffices if you are no longer using x afterwards. We do use it again in the unit tests, though. Bitcasting is defined in terms of memcpy, which then all compilers in practice elide, so that will never have this issue. Btw, I confirmed the issue to work using our Dockerfile, slightly modified to build and run vast-test rather than vast. I reproduced the issue on msater, and was unable to with this fix.
gharchive/pull-request
2022-12-19T22:15:28
2025-04-01T06:45:58.186441
{ "authors": [ "dominiklohmann" ], "repo": "tenzir/vast", "url": "https://github.com/tenzir/vast/pull/2814", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1154887732
fix: Errors on ignore_desired_capacity_changes = true with mixed_instances_policy Description fix: Errors on ignore_desired_capacity_changes = true with mixed_instances_policy Motivation and Context Configured the mixed_instances_policy object just like the example shows, but encounterd the following errors while applying an autoscaling group with ignore_desired_capacity_changes = true. ``` Error: Invalid function argument on .terraform/modules/asg/main.tf line 468, in resource "aws_autoscaling_group" "idc": 468: instance_type = lookup(override.value, "instance_type", null) ├──────────────── │ override.value is tuple with 6 elements Invalid value for "inputMap" parameter: lookup() requires a map as the first argument. Error: Invalid function argument on .terraform/modules/asg/main.tf line 469, in resource "aws_autoscaling_group" "idc": 469: weighted_capacity = lookup(override.value, "weighted_capacity", null) ├──────────────── │ override.value is tuple with 6 elements Invalid value for "inputMap" parameter: lookup() requires a map as the first argument. ``` If you set ignore_desired_capacity_changes = false with the same mixed_instance_policy, then terraform apply does not report these errors. Breaking Changes There are no breaking changes. It is a bug fix. How Has This Been Tested? [x] I have tested and validated these changes using the mixed_instance module in the provided examples/complete project (with ignore_desired_capacity_changes = true set) Tested with Terraform v1.1.4 @bryantbiggs This MR is not proceeding for 2weeks. Can you suggest me any actions missing from me? @smaruy30 we'll need a reproduction showing the issue and then that will most likely get added as a test case so we can validate and sure functionality is maintained going forward This issue has been resolved in version 6.0.0 :tada: @antonbabenko Thanks for the update!
gharchive/pull-request
2022-03-01T05:23:53
2025-04-01T06:45:58.326379
{ "authors": [ "antonbabenko", "bryantbiggs", "smaruy30" ], "repo": "terraform-aws-modules/terraform-aws-autoscaling", "url": "https://github.com/terraform-aws-modules/terraform-aws-autoscaling/pull/183", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2524383883
[add]: multiple service connect services Description I changed a line to allow the dynamic block to take an entier list instead of one element Motivation and Context Breaking Changes How Has This Been Tested? [ ] I have updated at least one of the examples/* to demonstrate and validate my change(s) [x] I have tested and validated these changes using one or more of the provided examples/* projects [ ] I have executed pre-commit run -a on my pull request closing in favor of #217
gharchive/pull-request
2024-09-13T09:41:02
2025-04-01T06:45:58.329497
{ "authors": [ "bryantbiggs", "mathias-mrsn" ], "repo": "terraform-aws-modules/terraform-aws-ecs", "url": "https://github.com/terraform-aws-modules/terraform-aws-ecs/pull/229", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
731626834
Usage Example breaks during cluster upgrades I have issues I'm submitting a... [x] bug report [ ] feature request [ ] support request - read the FAQ first! [ ] kudos, thank you, warm fuzzy What is the current behavior? Resources that use the Kubernetes provider from the example (including the aws-auth configmap in the module) will fail to be modified after a control plane upgrade with the following (or similar) error: Error: Failed to update Config Map: Unauthorized Right now, the recommended/example way to set use this module per the README is as follows: data "aws_eks_cluster" "cluster" { name = module.my-cluster.cluster_id } data "aws_eks_cluster_auth" "cluster" { name = module.my-cluster.cluster_id } provider "kubernetes" { host = data.aws_eks_cluster.cluster.endpoint cluster_ca_certificate = base64decode(data.aws_eks_cluster.cluster.certificate_authority.0.data) token = data.aws_eks_cluster_auth.cluster.token load_config_file = false version = "~> 1.9" } module "my-cluster" { source = "terraform-aws-modules/eks/aws" cluster_name = "my-cluster" cluster_version = "1.17" } Note that the two data sources are dependent on outputs from the module. During the first creation of the module, the dependency chain will look something like this: After the first creation of the module, that output is well-known and does not change during a cluster upgrade, so it seems like Terraform will allow the data sources to be read/updated before the cluster upgrade happens. That results in a graph that looks more like the following The problem that seems to be twofold: When a cluster upgrade does occur, it looks like tokens generated against the old control plane are no longer valid. I tested this by requesting and storing a token locally, then modifying a copy of the kubeconfig to use it explicitly. With time still left before the expiration, I received You must be logged in to the server (Unauthorized) when trying to access the cluster after the control plane was upgraded. I'm not sure if this is a bug with EKS/AWS. Since a cluster upgrade takes an extraordinarily long time (30m+), the token that was fetched in the datasource would have expired since they seem to have a lifetime of ~15m. That means that even in the absence of the problem described above, modifying the config map would still fail. If this is a bug, how to reproduce? Please include a code sample if relevant. Create a kubernetes cluster following the example, then upgrade the control plane while modifying the aws-auth configmap to add a role. Observe that the configmap modification fails with the following error: Error: Failed to update Config Map: Unauthorized What's the expected behavior? The ConfigMap should be updated. Are you able to fix this problem and submit a PR? Link here if you have already. Environment details Affected module version: 12.1.0 OS: OSX Terraform version: 0.12.29 Any other relevant info My current workaround that I just implemented is the following, though it creates a dependency on the aws cli: provider "kubernetes" { host = data.aws_eks_cluster.cluster.endpoint cluster_ca_certificate = base64decode(data.aws_eks_cluster.cluster.certificate_authority.0.data) load_config_file = false exec { api_version = "client.authentication.k8s.io/v1alpha1" command = "aws" args = ["eks", "get-token", "--cluster-name", "${module.eks.cluster_id}"] } } @dcherman I have the same issue and it seems that after update some autoscaling node groups cannot connect to the cluster due to authorization error (((. Also terraform is trying to create new aws_auth kubernetes resource. It happens always after update, i am going to increase duration_seconds for cluster IAM roles but I am not sure , that the problem is here( @dzirg44 Are you using managed nodegroups? If so, I've experienced similar problems. Are you experiencing this with the latest version of this module or with v12.2.0 ? @barryib In my case, v12.1.0, however in looking at the example I believe the latest version would have this issue as well. I haven't upgraded yet since I think I had a problem w/ cycles last time I tried; haven't revisited that yet. In v12.2.0 we improved dependences management for managed node group https://github.com/terraform-aws-modules/terraform-aws-eks/pull/867. I was wondering if this solves your issue.
gharchive/issue
2020-10-28T17:01:02
2025-04-01T06:45:58.340665
{ "authors": [ "barryib", "dcherman", "dzirg44" ], "repo": "terraform-aws-modules/terraform-aws-eks", "url": "https://github.com/terraform-aws-modules/terraform-aws-eks/issues/1072", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2525490203
feat: Allow passing in custom instance profile role Description Allows passing in a custom instance profile role and custom autoscaling role to the service pass role policy. Also includes minor provider version upgrade https://github.com/hashicorp/terraform-provider-aws/pull/37740. Motivation and Context Closes: https://github.com/terraform-aws-modules/terraform-aws-emr/issues/26 Breaking Changes No. How Has This Been Tested? [x] I have updated at least one of the examples/* to demonstrate and validate my change(s) [x] I have tested and validated these changes using one or more of the provided examples/* projects [x] I have executed pre-commit run -a on my pull request This PR is included in version 2.3.0 :tada:
gharchive/pull-request
2024-09-13T19:07:38
2025-04-01T06:45:58.344959
{ "authors": [ "antonbabenko", "magreenbaum" ], "repo": "terraform-aws-modules/terraform-aws-emr", "url": "https://github.com/terraform-aws-modules/terraform-aws-emr/pull/30", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1331740645
Wrong url has been occurred while running terraform apply. ╷ │ Error: failed to create kubernetes rest client for read of resource: Get "https://localhost/api?timeout=32s": x509: certificate is valid for *.dev.tappy.stage-codal.net, dev.tappy.stage-codal.net, not localhost │ │ with module.cert_manager.kubectl_manifest.cluster_issuer[0], │ on .terraform/modules/cert_manager/main.tf line 42, in resource "kubectl_manifest" "cluster_issuer": │ 42: resource "kubectl_manifest" "cluster_issuer" { │ Working fine in local machine where kubectl, helm, kubectl config and providers are configured. Not working in jenkins server where only kubectl command is installed. Hello, terraform providers settings are incorrect Hello, terraform providers settings are incorrect Kubernetes and helm resources through terraform are being created. Just this one throws error What about kubectl provider ? https://github.com/terraform-iaac/terraform-kubernetes-cert-manager#you-should-to-add-into-your-terraform-kubectl--helm--provider-configuration jenkins@ip-172-31-11-15:~/workspace/Dev-Terraform-Deployment/infra/terraform-resources/configs$ cat kubernetes.tf | head -n25 provider "helm" { kubernetes { host = module.eks.cluster_endpoint cluster_ca_certificate = base64decode(module.eks.cluster_certificate_authority_data) token = data.aws_eks_cluster_auth.eks_cluster.token } experiments { manifest = true } } provider "kubernetes" { host = module.eks.cluster_endpoint cluster_ca_certificate = base64decode(module.eks.cluster_certificate_authority_data) token = data.aws_eks_cluster_auth.eks_cluster.token } provider "kubectl" { host = module.eks.cluster_endpoint cluster_ca_certificate = base64decode(module.eks.cluster_certificate_authority_data) token = data.aws_eks_cluster_auth.eks_cluster.token } pls chk above I cannot understand how it is getting "https://localhost/api?timeout=32s"? please try to add this into kubectl provider provider "kubectl" { host = module.eks.cluster_endpoint cluster_ca_certificate = base64decode(module.eks.cluster_certificate_authority_data) token = data.aws_eks_cluster_auth.eks_cluster.token load_config_file = false } @bohdantverdyi Thank you. It has been resolved.
gharchive/issue
2022-08-08T11:35:43
2025-04-01T06:45:58.377396
{ "authors": [ "bohdantverdyi", "codal-jayesh" ], "repo": "terraform-iaac/terraform-kubernetes-cert-manager", "url": "https://github.com/terraform-iaac/terraform-kubernetes-cert-manager/issues/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1800693323
fix(deps): update dependency gitpython to v3.1.32 This PR contains the following updates: Package Update Change gitpython patch ==3.1.31 -> ==3.1.32 Release Notes gitpython-developers/GitPython (gitpython) v3.1.32: - with another security update Compare Source What's Changed Bump cygwin/cygwin-install-action from 3 to 4 by @​dependabot in https://github.com/gitpython-developers/GitPython/pull/1572 Fix up the commit trailers functionality by @​itsluketwist in https://github.com/gitpython-developers/GitPython/pull/1576 Name top-level exceptions as private variables by @​Hawk777 in https://github.com/gitpython-developers/GitPython/pull/1590 fix pypi long description by @​eUgEntOptIc44 in https://github.com/gitpython-developers/GitPython/pull/1603 Don't rely on del by @​r-darwish in https://github.com/gitpython-developers/GitPython/pull/1606 Block insecure non-multi options in clone/clone_from by @​Beuc in https://github.com/gitpython-developers/GitPython/pull/1609 New Contributors @​Hawk777 made their first contribution in https://github.com/gitpython-developers/GitPython/pull/1590 @​eUgEntOptIc44 made their first contribution in https://github.com/gitpython-developers/GitPython/pull/1603 @​r-darwish made their first contribution in https://github.com/gitpython-developers/GitPython/pull/1606 @​Beuc made their first contribution in https://github.com/gitpython-developers/GitPython/pull/1609 Full Changelog: https://github.com/gitpython-developers/GitPython/compare/3.1.31...3.1.32 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 Ignore: Close this PR and you won't be reminded about this update again. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline
gharchive/pull-request
2023-07-12T10:40:48
2025-04-01T06:45:58.391074
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/common-dev-assets", "url": "https://github.com/terraform-ibm-modules/common-dev-assets/pull/681", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2128093053
chore(deps): update common-dev-assets digest to 3324eff This PR contains the following updates: Package Update Change common-dev-assets digest ce80a6d -> 3324eff Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline :tada: This PR is included in version 1.0.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-02-10T04:43:03
2025-04-01T06:45:58.401608
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-enterprise", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-enterprise/pull/137", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2498556996
chore(deps): update ci dependencies This PR contains the following updates: Package Type Update Change common-dev-assets digest 0b1411f -> 5b96dd7 github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper require minor v1.37.2 -> v1.38.1 Release Notes terraform-ibm-modules/ibmcloud-terratest-wrapper (github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper) v1.38.1 Compare Source Bug Fixes deps: update gomod (#​857) (87710ff) v1.38.0 Compare Source Features add function to check ingress health status of a cluster (#​751) - Added new function testhelper.CheckClusterIngressHealthy() which can be used in Terraform tests to validate the status of an OpenShift cluster (801dd27) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline ℹ Artifact update notice File name: tests/go.mod In order to perform the update(s) described in the table above, Renovate ran the go get command, which resulted in the following additional change(s): 3 additional dependencies were updated Details: Package Change github.com/IBM/platform-services-go-sdk v0.67.0 -> v0.68.1 github.com/gruntwork-io/terratest v0.47.0 -> v0.47.1 github.com/hashicorp/go-getter v1.7.5 -> v1.7.6 /run pipeline :tada: This PR is included in version 1.1.1 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-08-31T02:12:01
2025-04-01T06:45:58.415932
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-iam-serviceid-apikey-secrets-manager", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-iam-serviceid-apikey-secrets-manager/pull/144", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2290489016
fix(deps): update terraform ibm to latest for the deployable architecture solution This PR contains the following updates: Package Type Update Change ibm (source) required_provider minor 1.64.2 -> 1.65.0 Release Notes IBM-Cloud/terraform-provider-ibm (ibm) v1.65.0 Compare Source Features Support for VMware as a Service API Datasources ibm_vmaas_vdc Resources ibm_vmaas_vdc Support for PowerVS Datasources ibm_pi_networks Enhancements align Projects terraform provider to latest API spec (5252) Add power edge details to workspace data source (5205) feat(tekton): add locked properties (5274) Added support for IBM Cloud Logs atracker targets (5270) Deprecate vtl attributes in data source/resource (5265) Update PI Image import/export and capture cos region doc (5273) CBR typeset change for cbr rule/zone and adding CBR retries (5246) Fix data assignment APIs (5306) feat: wrap provider to create new error structure (5278) ODF workerpool support (5283) fix(Cloud Databases):Incorporate new client-error-toolchain in datasources (5320) Added IAM fields in enterprise create child account response (5313) ODF 4.15 Templates for add-on and satellite (5316) Added support for snapshot service tags (5326) BugFixes Fix deleting instances with replicant attribute set (5176) fix access group resource: recreate ag when it is deleted outside terraform (5304) fix(docs): fixed documents for network acl and rules (5302) add fix for timeouts for observe logging resource (5300) Produce error when selecting invalid region in toolchain service (5297) Fix instance policy drift (5308) fix access group members resource: recreate ag members when it is deleted outside terraform (5317) updated host_flavor id docs (5323) fix(docs): Fix bm initialization doc titile (5330) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline /run pipeline /run pipeline :tada: This PR is included in version 1.10.9 :tada: The release is available on: GitHub release v1.10.9 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-05-10T22:33:34
2025-04-01T06:45:58.436725
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-icd-elasticsearch", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-icd-elasticsearch/pull/157", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2663487691
chore(deps): update ci dependencies This PR contains the following updates: Package Type Update Change common-dev-assets digest 95f6c1e -> 81754c2 github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper require patch v1.41.4 -> v1.41.5 Release Notes terraform-ibm-modules/ibmcloud-terratest-wrapper (github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper) v1.41.5 Compare Source Bug Fixes deps: update module github.com/ibm/vpc-go-sdk to v0.62.0 (#​896) (f84c9e7) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline ℹ Artifact update notice File name: tests/go.mod In order to perform the update(s) described in the table above, Renovate ran the go get command, which resulted in the following additional change(s): 4 additional dependencies were updated The go directive was updated for compatibility reasons Details: Package Change go 1.22.0 -> 1.22.4 github.com/IBM/cloud-databases-go-sdk v0.7.0 -> v0.7.1 github.com/IBM/platform-services-go-sdk v0.71.0 -> v0.71.1 github.com/IBM/schematics-go-sdk v0.3.0 -> v0.4.0 golang.org/x/net v0.25.0 -> v0.30.0 /run pipeline ℹ Artifact update notice File name: tests/go.mod In order to perform the update(s) described in the table above, Renovate ran the go get command, which resulted in the following additional change(s): 4 additional dependencies were updated The go directive was updated for compatibility reasons Details: Package Change go 1.22.0 -> 1.22.4 github.com/IBM/cloud-databases-go-sdk v0.7.0 -> v0.7.1 github.com/IBM/platform-services-go-sdk v0.71.0 -> v0.71.1 github.com/IBM/schematics-go-sdk v0.3.0 -> v0.4.0 golang.org/x/net v0.25.0 -> v0.30.0 /run pipeline ℹ Artifact update notice File name: tests/go.mod In order to perform the update(s) described in the table above, Renovate ran the go get command, which resulted in the following additional change(s): 4 additional dependencies were updated The go directive was updated for compatibility reasons Details: Package Change go 1.22.0 -> 1.22.4 github.com/IBM/cloud-databases-go-sdk v0.7.0 -> v0.7.1 github.com/IBM/platform-services-go-sdk v0.71.0 -> v0.71.1 github.com/IBM/schematics-go-sdk v0.3.0 -> v0.4.0 golang.org/x/net v0.25.0 -> v0.30.0 /run pipeline /run pipeline :tada: This PR is included in version 3.19.10 :tada: The release is available on: GitHub release v3.19.10 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-11-15T23:39:39
2025-04-01T06:45:58.463449
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-icd-postgresql", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-icd-postgresql/pull/528", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1638784985
chore(deps): update common-dev-assets digest to 8039123 This PR contains the following updates: Package Update Change common-dev-assets digest e9138b0 -> 8039123 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. :tada: This PR is included in version 1.3.3 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-03-24T06:13:14
2025-04-01T06:45:58.469359
{ "authors": [ "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-icd-postgresql", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-icd-postgresql/pull/98", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2034463873
chore(deps): update common-dev-assets digest to dfefa07 This PR contains the following updates: Package Update Change common-dev-assets digest aab307f -> dfefa07 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline :tada: This PR is included in version 2.3.2 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-12-10T17:35:36
2025-04-01T06:45:58.475012
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-kms-key-ring", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-kms-key-ring/pull/501", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2277447930
Customise some input variables in ibm_catalog.json Review all your input variables in https://github.com/terraform-ibm-modules/terraform-ibm-mas/blob/main/solutions/existing-cluster/variables.tf and for any that can be customised (e.g. dropdown pickers) please add them into the ibm_catalog.json with the customisations. As an example, you might wish to convert the region input variable into a dropdown like was done for landing zone here. And maybe add a custom dropdown list for deployment_flavour that only shows core and manage in the dropdown. For example of custom dropdown, see this. Re-opening, the PR is now yet merged PR is now merged - closing
gharchive/issue
2024-05-03T11:05:17
2025-04-01T06:45:58.478243
{ "authors": [ "ocofaigh" ], "repo": "terraform-ibm-modules/terraform-ibm-mas", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-mas/issues/84", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1772536041
chore(deps): update common-dev-assets digest to 0b2bca3 This PR contains the following updates: Package Update Change common-dev-assets digest c6b4a5b -> 0b2bca3 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline /run pipeline /run pipeline Edited/Blocked Notification Renovate will not automatically rebase this PR, because it does not recognize the last commit author and assumes somebody else may have edited the PR. You can manually request rebase by checking the rebase/retry box above. ⚠ Warning: custom changes will be lost. :tada: This PR is included in version 1.2.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-06-24T08:29:49
2025-04-01T06:45:58.485350
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-module-template", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-module-template/pull/582", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2229489834
chore(deps): update common-dev-assets digest to 2285d99 This PR contains the following updates: Package Update Change common-dev-assets digest 0bb740c -> 2285d99 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline :tada: This PR is included in version 1.1.6 :tada: The release is available on: GitHub release v1.1.6 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-04-07T00:48:40
2025-04-01T06:45:58.491851
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-observability-da", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-observability-da/pull/30", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1761648181
chore(deps): update module github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper to v1.10.5 This PR contains the following updates: Package Type Update Change github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper require patch v1.10.4 -> v1.10.5 Release Notes terraform-ibm-modules/ibmcloud-terratest-wrapper v1.10.5 Compare Source Bug Fixes deps: update module golang.org/x/crypto to v0.10.0 (#​566) (5e52db4) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline :tada: This PR is included in version 6.2.5 :tada: The release is available on: GitHub release v6.2.5 Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-06-17T05:09:38
2025-04-01T06:45:58.501200
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-powervs-sap", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-powervs-sap/pull/408", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2447281454
fix(deps): update terraform ibm to latest for the deployable architecture solution This PR contains the following updates: Package Type Update Change ibm (source) required_provider minor 1.67.1 -> 1.68.0 Release Notes IBM-Cloud/terraform-provider-ibm (ibm) v1.68.0 Compare Source Features Support CBR Datasources ibm_cbr_zone_addresses Resources ibm_cbr_zone_addresses Support CIS Datasources ibm_cis_origin_certificates Resources ibm_cis_advanced_certificate_pack_order ibm_cis_origin_certificate_order Support VPC Resources ibm_is_bare_metal_server_initialization Enhancements Add dhcp network support for stratos (5503) IAM Policy Assignment: S2S Policy Assignments (5499) Code Engine Provider and Documentation Update (5347) move wait_till logic into function, integrate it into vpc_cluster datasource (5476) Remove hardcoded values for private and direct cos config endpoint (5484) feat(bm-dynamic-bandwidth): Support for bandwidth in bare metal (5493) Doc update for ODF (5454) feat(fs-cross-account): Support for file share cross account access (5510) feat Bm firmware update (5519) Changing the documentation for SCC (5456) feat(lb-parameterized-redirect): Update doc to specify parameterized url redirect (5521) chore(Cloud-Databases): Remove Datastax (5511) feat(ResourceController): Added onetime_credentials to ResourceInstance and ResourceKey read schemas (5532) Add PhysicalAddress and CapabilitiesManagedBySatellite to Terraform SatelliteLocation (5530) Added support for bm reinitialization (5520) BugFixes fix(CIS): updating managed ruleset documents (5488) fix(ins-keys): Make VSI keys optional (5518) fix(is-volume): Set catalogoffering computed attribute empty list (5514) add import_on_create param to ibm_container_vpc_worker_pool doc (5506) [Doc]Update Doc for E1080 Support (5536) Fix(iam-identity):trusted profile templates (5440) S2S documentation note for ibm_kms_key and minor linting fixes (5529) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline :tada: This issue has been resolved in version 1.15.1 :tada: The release is available on: GitHub release v1.15.1 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-08-04T20:14:59
2025-04-01T06:45:58.523363
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-scc-da", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-scc-da/pull/158", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2229496346
chore(deps): update common-dev-assets digest to 2285d99 This PR contains the following updates: Package Update Change common-dev-assets digest 0bb740c -> 2285d99 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline :tada: This PR is included in version 1.3.0 :tada: The release is available on: GitHub release v1.3.0 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-04-07T01:14:21
2025-04-01T06:45:58.529867
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-scc-da", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-scc-da/pull/50", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2739427643
fix(deps): update required_provider to latest for deployable architectures This PR contains the following updates: Package Type Update Change ibm (source) required_provider minor 1.71.3 -> 1.72.2 kubernetes (source) required_provider minor 2.33.0 -> 2.35.0 Release Notes IBM-Cloud/terraform-provider-ibm (ibm) v1.72.2 Compare Source BugFixes Fix the issue of loading IAM endpoints either for stage or file (5866) v1.72.1 Compare Source BugFixes Breaking change fix Revert "Add version validation (5859) v1.72.0 Compare Source Features Support for Db2 SaaS Resources ibm_db2 Datasources ibm_db2 Support for IAM Datasources ibm_iam_effective_account_settings Support for MQaaS -Resources - ibm_mqcloud_virtual_private_endpoint_gateway Datasources ibm_mqcloud_virtual_private_endpoint_gateway ibm_mqcloud_virtual_private_endpoint_gateways Support for VPC Datasources ibm_is_cluster_network ibm_is_cluster_networks ibm_is_cluster_network_interface ibm_is_cluster_network_interfaces ibm_is_cluster_network_profile ibm_is_cluster_network_profiles ibm_is_cluster_network_subnet ibm_is_cluster_network_subnets ibm_is_cluster_network_subnet_reserved_ip ibm_is_cluster_network_subnet_reserved_ips ibm_is_instance_cluster_network_attachment ibm_is_instance_cluster_network_attachments Resources ibm_is_cluster_network_interface ibm_is_cluster_network_subnet_reserved_ip ibm_is_cluster_network_subnet ibm_is_cluster_network ibm_is_instance_cluster_network_attachment Support for Power System Datasources ibm_pi_network_peers Enhancements Add version validation (5791) Refactor shared-processor-pool resource (5796) Refactor Cloud Connection Network Attach (5781) Import a resource provisioned with key_protect_id and key_protect_instance attributes set (5798) enable commit while creating template/version (5797) ListClassicWorkers => ListAllWorkers (5784) chore: update mirroring example doc (5767) Update resource_instance.html.markdown (5806) fix(cloud-databases): Refactor tests Part 1 (5810) feat(BM-Reservations-Automatic-Attachment): Baremetal Server Reservations and Reservations Automatic Attachment (5805) Support enable secure-by-default (5751) Added support for volume_prototypes on ibm_is_instance resource (5777) ODF 4.16 terraform support (5789) Add Chenai to the doc (5813) ibm_pag_instance should force a new instance when parameters_json are updated (5807) Add power private datacenter capabilities (5833) fixed instance, instance template documents, alignment (5850) fix: instance cluster network attachment update (5853) Custom Diff User Tags (5844) added primary_ip refernce to nac for consistency (5847) BugFixes ICD: Promote read replicas (5738) docs(pps): Remove SA notes fpr private path service gateway feature (5788) Fix sap profile datasource warning (5779) fixed documents for is datasources (5782) fix: eventstreams sarama clientid (5802) fix(ibm_is_image): 404 error fix on datasource (5769) fix(ibm_is_share): 404 error fix on datasource (5770) fix(ibm_is_snapshot): 404 error fix on datasource (5771) vol fix for snapshot crn (5815) fix: adjust validation for Code Engine application scale_concurrency_target (5792) hashicorp/terraform-provider-kubernetes (kubernetes) v2.35.0 Compare Source FEATURES: resources_kubernetes_daemon_set_v1 : Added max_surge argument for to rolling_update block. [GH-2630] v2.34.0 Compare Source ENHANCEMENTS: Added conditions attribute to kubernetes_nodes data source, which will provide detailed node health and status information [GH-2612] Adding the kubernetes_secret_v1_data resource to the kubernetes provider. This resource will allow users to manage kubernetes secrets [GH-2604] Properly handle Kubernetes Jobs with ttl_seconds_after_finished = 0 to prevent unnecessary recreation. [GH-2596] FEATURES: New ephemeral resource: kubernetes_certificate_signing_request_v1 [GH-2628] New ephemeral resource: kubernetes_token_request_v1 [GH-2628] Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR is behind base branch, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline /run pipeline :tada: This PR is included in version 1.4.2 :tada: The release is available on: GitHub release v1.4.2 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-12-14T01:14:12
2025-04-01T06:45:58.571054
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-scc-workload-protection-agent", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-scc-workload-protection-agent/pull/222", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2746415045
chore(deps): update common-dev-assets digest to 5f5fd35 This PR contains the following updates: Package Update Change common-dev-assets digest d63f9de -> 5f5fd35 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR is behind base branch, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline
gharchive/pull-request
2024-12-18T00:27:08
2025-04-01T06:45:58.576458
{ "authors": [ "terraform-ibm-modules-dev", "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-secrets-manager-public-cert", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-secrets-manager-public-cert/pull/140", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1679258847
chore(deps): update module github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper to v1.8.9 This PR contains the following updates: Package Type Update Change github.com/terraform-ibm-modules/ibmcloud-terratest-wrapper require patch v1.8.8 -> v1.8.9 Release Notes terraform-ibm-modules/ibmcloud-terratest-wrapper v1.8.9 Compare Source Bug Fixes deps: update module github.com/ibm/platform-services-go-sdk to v0.36.0 (#​528) (f4a94b5) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. :tada: This PR is included in version 1.0.1 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-04-22T01:27:58
2025-04-01T06:45:58.585320
{ "authors": [ "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-security-group", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-security-group/pull/62", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1619791709
chore(deps): update common-dev-assets digest to 04de466 This PR contains the following updates: Package Update Change common-dev-assets digest c1e8ace -> 04de466 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. :tada: This PR is included in version 2.1.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-03-11T00:22:10
2025-04-01T06:45:58.590997
{ "authors": [ "terraform-ibm-modules-ops" ], "repo": "terraform-ibm-modules/terraform-ibm-transit-gateway", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-transit-gateway/pull/205", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2512085749
chore(deps): update terraform-module This PR contains the following updates: Package Type Update Change github.com/terraform-ibm-modules/terraform-ibm-landing-zone module patch v5.31.1 -> v5.31.2 terraform-ibm-modules/base-ocp-vpc/ibm (source) module patch 3.30.0 -> 3.30.1 Release Notes terraform-ibm-modules/terraform-ibm-landing-zone (github.com/terraform-ibm-modules/terraform-ibm-landing-zone) v5.31.2 Compare Source Bug Fixes added outputs for workload/management cluster names (#​875) (a6b3f6e) terraform-ibm-modules/terraform-ibm-base-ocp-vpc (terraform-ibm-modules/base-ocp-vpc/ibm) v3.30.1 Compare Source Bug Fixes deps: update terraform-module (#​512) (3aecbb6) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline /run pipeline @ocofaigh we need to add the same change on TestRunSLZExample test adding it /run pipeline :tada: This PR is included in version 1.4.25 :tada: The release is available on: GitHub release v1.4.25 Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-09-07T21:33:47
2025-04-01T06:45:58.605176
{ "authors": [ "ocofaigh", "terraform-ibm-modules-dev", "terraform-ibm-modules-ops", "vbontempi" ], "repo": "terraform-ibm-modules/terraform-ibm-websphere-liberty-operator", "url": "https://github.com/terraform-ibm-modules/terraform-ibm-websphere-liberty-operator/pull/126", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
466404712
Kayrus optional parameters Addressing #230, #205, and #190 where GITHUB_TOKEN should not be required for all data sources. Before fix: Error: GET https://api.github.com/users/hashibot: 401 Bad credentials [] on /var/folders/nj/n5hr70hd2dx45qb__6wl10nc0000gp/T/tf-test353694711/main.tf line 6: (source code not available) FAIL After fix: == RUN TestProvider_anonymous --- PASS: TestProvider_anonymous (1.04s) PASS @megan07 when are you going to make the release with this feature?
gharchive/pull-request
2019-07-10T16:24:55
2025-04-01T06:45:58.689090
{ "authors": [ "kayrus", "megan07" ], "repo": "terraform-providers/terraform-provider-github", "url": "https://github.com/terraform-providers/terraform-provider-github/pull/255", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2064240440
V1.5.0 release changes Update Changelog Change Examples links to use v1.5.0 tag /lgtm /approve
gharchive/pull-request
2024-01-03T15:38:45
2025-04-01T06:45:58.740086
{ "authors": [ "bardielle", "nirarg" ], "repo": "terraform-redhat/terraform-provider-rhcs", "url": "https://github.com/terraform-redhat/terraform-provider-rhcs/pull/487", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1045438669
Remove file name checks when reading in RIPL data. Remove check for number of files and file names when reading the RIPL data into the nuclide bases. This check was found to be incorrect for the latest version of RIPL (from 7/12/2020) where the file names were no longer labeled as z000.dat to z117.dat. Instead they were changed to z001.dat to z118.dat. On balance, it seems like a clear, easy-to-understand change. The obvious questions I have are: Now that we have removed this check, what if one or more files are missing? Will the user know why their run failed? Why not add in a simpler check to make sure that there are >= 118 files named *.dat? (That does mostly the same thing, but is a bit more future proof.) Hi @john-science - These are good questions and I will add the responses here on the PR so we can track them: The original implementation was to make sure that this was checked, but the way that this functions now is that the nuclideBases will be populated based on the zXXX.dat files. There are missing zXXX.dat files then some of the nuclideBases may not be initialized and therefore the if some client code needs that nuclide then a KeyError exception will be thrown. We could check the number of files, but that may not be future proof. I could add at least 118 files possibly, if that helps?
gharchive/pull-request
2021-11-05T04:16:21
2025-04-01T06:45:58.752975
{ "authors": [ "jakehader", "john-science" ], "repo": "terrapower/armi", "url": "https://github.com/terrapower/armi/pull/462", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1608403614
default threads=1? currently 8, but that's probably not an appropriate default. Pros: avoids users needing to remember to set the number of threads; in most cases, we can expect that they'll want multi-threaded indexing and retrieval nearly free speedup Cons: Easy for folks to mistakenly measure query throughput, rather than latency 8 threads will rarely be optimal, machines will often have more threads that should be used. Or users may have multiple processes going at once without realising that Pisa is multi-threaded, causing thrashing fixed
gharchive/issue
2023-03-03T10:46:42
2025-04-01T06:45:58.756280
{ "authors": [ "seanmacavaney" ], "repo": "terrierteam/pyterrier_pisa", "url": "https://github.com/terrierteam/pyterrier_pisa/issues/17", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2104419923
똥피하기게임_신준원 A,D / <-, -> - 좌우 이동 스페이스바 - 총 쏘기 설명 추가
gharchive/pull-request
2024-01-29T00:18:09
2025-04-01T06:45:58.778285
{ "authors": [ "tesia102" ], "repo": "tesia102/Avoid_Ddong_And_Shoot", "url": "https://github.com/tesia102/Avoid_Ddong_And_Shoot/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1480024461
🛑 racl.kr is down In f5d9781, racl.kr (https://racl.kr) was down: HTTP code: 500 Response time: 3265 ms Resolved: racl.kr is back up in 9a25d27.
gharchive/issue
2022-12-06T19:45:42
2025-04-01T06:45:58.782980
{ "authors": [ "teslamint" ], "repo": "teslamint/uptime", "url": "https://github.com/teslamint/uptime/issues/179", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
468028488
Adding confidential threshold on output Where can I obtain the degree of confidence that every word tesseract recognized? I'm dealing with important messages that would better be left as blank than mistakenly translated. I want to set up a threshold to let tesseract output "?" if the degree of confidence for a given word/character is low. Is it possible? Environment Tesseract 4.0 ubuntu 64-bit Please respect guidelines for posting issue: use tesseract user forum for asking questions/support.
gharchive/issue
2019-07-15T09:32:30
2025-04-01T06:45:58.788550
{ "authors": [ "zdenop", "zguoguo" ], "repo": "tesseract-ocr/tesseract", "url": "https://github.com/tesseract-ocr/tesseract/issues/2566", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2397758314
[Bug]: Finalizing JUnit 5 test case with Kafka Streams and Spring Boot takes several minutes Module Kafka Testcontainers version 1.19.8 Using the latest Testcontainers version? Yes Host OS MacOS, Linux, Windows Host Arch ARM, x86 Docker version This is my local setup with Podman. In the pipeline we use Docker and other colleagues use Docker Desktop where the issue also happens. Client: Docker Engine - Community Version: 27.0.3 API version: 1.41 (downgraded from 1.46) Go version: go1.22.4 Git commit: 7d4bcd863a Built: Fri Jun 28 14:56:30 2024 OS/Arch: darwin/arm64 Context: default Server: linux/arm64/fedora-40 Podman Engine: Version: 5.0.3 APIVersion: 5.0.3 Arch: arm64 BuildTime: 2024-05-10T02:00:00+02:00 Experimental: false GitCommit: GoVersion: go1.22.2 KernelVersion: 6.8.8-300.fc40.aarch64 MinAPIVersion: 4.0.0 Os: linux Conmon: Version: conmon version 2.1.10, commit: Package: conmon-2.1.10-1.fc40.aarch64 OCI Runtime (crun): Version: crun version 1.14.4 commit: a220ca661ce078f2c37b38c92e66cf66c012d9c1 rundir: /run/crun spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +LIBKRUN +WASM:wasmedge +YAJL Package: crun-1.14.4-1.fc40.aarch64 Engine: Version: 5.0.3 API version: 1.41 (minimum version 1.24) Go version: go1.22.2 Git commit: Built: Fri May 10 02:00:00 2024 OS/Arch: linux/arm64 Experimental: false What happened? We're running a Spring Boot project with Kafka Streams. We have one integration test (using @SpringBootTest and @Testcontainers) that sometimes takes several minutes to complete after all tests successfully ran. Lots of Kafka "no connection to node" and "Node 1 disconnected" are printed in the logs repeatedly. It seems like the Kafka container is stopped before the Spring context is shutdown. And the Spring Context is waiting for Kafka to become available before shutting down. Relevant log output ### The following 20 lines are printed repeatedly for several minutes [11:00:35.147] [INFO] [org.apache.kafka.clients.NetworkClient] - [Consumer clientId=<our-consumer-group>-StreamThread-1-consumer, groupId=streams.notification-report-selector.notification-repartitioner] Node 1 disconnected. [11:00:35.147] [WARN] [org.apache.kafka.clients.NetworkClient] - [Consumer clientId=<our-consumer-group>-StreamThread-1-consumer, groupId=streams.notification-report-selector.notification-repartitioner] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:35.218] [INFO] [org.apache.kafka.clients.NetworkClient] - [AdminClient clientId=<our-consumer-group>-admin] Node 1 disconnected. [11:00:35.218] [WARN] [org.apache.kafka.clients.NetworkClient] - [AdminClient clientId=<our-consumer-group>-admin] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:35.350] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-3] Node 1 disconnected. [11:00:35.350] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-3] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:35.591] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=<our-consumer-group>-StreamThread-1-producer] Node 1 disconnected. [11:00:35.591] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=<our-consumer-group>-StreamThread-1-producer] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:35.782] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-1] Node 1 disconnected. [11:00:35.782] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-1] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.072] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-2] Node 1 disconnected. [11:00:36.072] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-2] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.148] [INFO] [org.apache.kafka.clients.NetworkClient] - [Consumer clientId=<our-consumer-group>-StreamThread-1-consumer, groupId=[<our-consumer-group>]] Node 1 disconnected. [11:00:36.148] [WARN] [org.apache.kafka.clients.NetworkClient] - [Consumer clientId=<our-consumer-group>-StreamThread-1-consumer, groupId=[<our-consumer-group>]] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.221] [INFO] [org.apache.kafka.clients.NetworkClient] - [AdminClient clientId=<our-consumer-group>-admin] Node 1 disconnected. [11:00:36.221] [WARN] [org.apache.kafka.clients.NetworkClient] - [AdminClient clientId=<our-consumer-group>-admin] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.370] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-3] Node 1 disconnected. [11:00:36.371] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-3] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.544] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=<our-consumer-group>-StreamThread-1-producer] Node 1 disconnected. [11:00:36.544] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=<our-consumer-group>-StreamThread-1-producer] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.731] [INFO] [org.apache.kafka.clients.admin.internals.AdminMetadataManager] - [AdminClient clientId=<our-consumer-group>-admin] Metadata update failed org.apache.kafka.common.errors.TimeoutException: Timed out waiting for a node assignment. Call: fetchMetadata [11:00:36.804] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-1] Node 1 disconnected. [11:00:36.804] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-1] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. [11:00:36.925] [INFO] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-2] Node 1 disconnected. [11:00:36.925] [WARN] [org.apache.kafka.clients.NetworkClient] - [Producer clientId=producer-2] Connection to node 1 (localhost/127.0.0.1:32772) could not be established. Node may not be available. ### The following log lines are only printed once at when the tests are finalized Error: 1:00:36.961] [ERROR] [org.apache.kafka.streams.processor.internals.TaskExecutor] - stream-thread [<our-consumer-group>-StreamThread-1] Committing task(s) 0_1 failed. org.apache.kafka.common.errors.TimeoutException: Timeout of 60000ms expired before successfully committing offsets {notification.managed-1=OffsetAndMetadata{offset=5, leaderEpoch=null, metadata='AgAAAZCWstBN'}} Error: 1:00:36.961] [ERROR] [org.apache.kafka.streams.processor.internals.TaskManager] - stream-thread [<our-consumer-group>-StreamThread-1] Exception caught while committing tasks [TaskId: 0_1 ProcessorTopology: Notifications: topics: [notification.managed] children: [Event-Type-Filter] Event-Type-Filter: children: [Key-Selector] Key-Selector: children: [Notifications-By-Id] Notifications-By-Id: topic: StaticTopicNameExtractor(notification.by-id) Partitions [notification.managed-1] ] org.apache.kafka.common.errors.TimeoutException: Timeout of 60000ms expired before successfully committing offsets {notification.managed-1=OffsetAndMetadata{offset=5, leaderEpoch=null, metadata='AgAAAZCWstBN'}} [11:00:36.961] [INFO] [org.apache.kafka.streams.processor.internals.StreamTask] - stream-thread [<our-consumer-group>-StreamThread-1] task [0_0] Suspended from RUNNING [11:00:36.962] [INFO] [org.apache.kafka.streams.processor.internals.RecordCollectorImpl] - stream-thread [<our-consumer-group>-StreamThread-1] stream-task [0_0] Closing record collector dirty [11:00:36.962] [INFO] [org.apache.kafka.streams.processor.internals.StreamTask] - stream-thread [<our-consumer-group>-StreamThread-1] task [0_0] Closed dirty [11:00:36.962] [INFO] [org.apache.kafka.streams.processor.internals.StreamTask] - stream-thread [<our-consumer-group>-StreamThread-1] task [0_1] Suspended from RUNNING [11:00:36.963] [INFO] [org.apache.kafka.streams.processor.internals.RecordCollectorImpl] - stream-thread [<our-consumer-group>-StreamThread-1] stream-task [0_1] Closing record collector dirty [11:00:36.963] [INFO] [org.apache.kafka.streams.processor.internals.StreamTask] - stream-thread [<our-consumer-group>-StreamThread-1] task [0_1] Closed dirty [11:00:36.963] [INFO] [org.apache.kafka.clients.producer.KafkaProducer] - [Producer clientId=<our-consumer-group>-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. [11:00:36.966] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics scheduler closed [11:00:36.966] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.metrics.JmxReporter [11:00:36.966] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.telemetry.internals.ClientTelemetryReporter [11:00:36.966] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics reporters closed [11:00:36.967] [INFO] [org.apache.kafka.common.utils.AppInfoParser] - App info kafka.producer for <our-consumer-group>-StreamThread-1-producer unregistered Error: 1:00:36.967] [ERROR] [org.apache.kafka.streams.processor.internals.StreamThread] - stream-thread [<our-consumer-group>-StreamThread-1] Failed to close task manager due to the following error: org.apache.kafka.common.errors.TimeoutException: Timeout of 60000ms expired before successfully committing offsets {notification.managed-1=OffsetAndMetadata{offset=5, leaderEpoch=null, metadata='AgAAAZCWstBN'}} [11:00:36.967] [INFO] [org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer] - [Consumer clientId=<our-consumer-group>-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions [11:00:36.967] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics scheduler closed [11:00:36.967] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.metrics.JmxReporter [11:00:36.967] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.telemetry.internals.ClientTelemetryReporter [11:00:36.967] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics reporters closed [11:00:36.970] [INFO] [org.apache.kafka.common.utils.AppInfoParser] - App info kafka.consumer for <our-consumer-group>-StreamThread-1-consumer unregistered [11:00:36.970] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics scheduler closed [11:00:36.970] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.metrics.JmxReporter [11:00:36.970] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.telemetry.internals.ClientTelemetryReporter [11:00:36.970] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics reporters closed [11:00:36.972] [INFO] [org.apache.kafka.common.utils.AppInfoParser] - App info kafka.consumer for <our-consumer-group>-StreamThread-1-restore-consumer unregistered [11:00:36.973] [INFO] [org.apache.kafka.streams.processor.internals.StreamThread] - stream-thread [<our-consumer-group>-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD [11:00:36.973] [INFO] [org.apache.kafka.streams.processor.internals.StreamThread] - stream-thread [<our-consumer-group>-StreamThread-1] Shutdown complete [11:00:36.973] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Shutdown 1 stream threads complete [11:00:36.973] [INFO] [org.apache.kafka.common.utils.AppInfoParser] - App info kafka.admin.client for <our-consumer-group>-admin unregistered [11:00:36.973] [INFO] [org.apache.kafka.clients.admin.internals.AdminMetadataManager] - [AdminClient clientId=<our-consumer-group>-admin] Metadata update failed org.apache.kafka.common.errors.TimeoutException: The AdminClient thread has exited. Call: fetchMetadata [11:00:36.973] [INFO] [org.apache.kafka.clients.admin.KafkaAdminClient] - [AdminClient clientId=<our-consumer-group>-admin] Timed out 1 remaining operation(s) during close. [11:00:36.984] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics scheduler closed [11:00:36.984] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.metrics.JmxReporter [11:00:36.984] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics reporters closed [11:00:36.984] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics scheduler closed [11:00:36.984] [INFO] [org.apache.kafka.common.metrics.Metrics] - Closing reporter org.apache.kafka.common.metrics.JmxReporter [11:00:36.984] [INFO] [org.apache.kafka.common.metrics.Metrics] - Metrics reporters closed [11:00:36.985] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] State transition from PENDING_SHUTDOWN to NOT_RUNNING [11:00:36.985] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client stopped completely [11:00:37.006] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal NOT_RUNNING state, all resources are closed and the client has stopped. [11:00:37.007] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal NOT_RUNNING state, all resources are closed and the client has stopped. [11:00:37.007] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal NOT_RUNNING state, all resources are closed and the client has stopped. [11:00:37.007] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal NOT_RUNNING state, all resources are closed and the client has stopped. [11:00:37.011] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.011] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.011] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.011] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.016] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.016] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.016] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.016] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.028] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.028] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.029] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.029] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.044] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.044] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.044] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. [11:00:37.044] [INFO] [org.apache.kafka.streams.KafkaStreams] - stream-client [<our-consumer-group>] Streams client is already in the terminal ERROR state, all resources are closed and the client has stopped. Additional Information The issue seems to be mitigated to some extend (we wait < 10 seconds for the test to finalize) if we start the container manually in the @BeforeAll method, instead of using the @Testcontainers and @Container annotations. Does anyone have a clue why this can happen? Or how I can investigate further? Hi, do you a project to reproduce it? I'll try to provide you with a minimal example today. Sorry for the slow response. Here is the example repo: https://github.com/MarvinVaillant/demo-long-kafka-streams-teardown/tree/main Here you'll also find a short video that shows the issue: https://github.com/user-attachments/assets/535847ac-cbbc-4f72-b6c9-a09e682bf84e @MarvinVaillant thanks for providing the example but I am not able to run the project. It fails because of Mapped port can only be obtained after the container is started. I can make some changes but I can not see the behavior described. i'm really sorry. I was fiddeling around with the code to get some more insights. Could you please pull the latest commits? Thanks for updating it. I'm moving this to a discussion because It is not an issue itself. Testcontainers is creating, running the container and killing it. However, in order to close the context you should call close explicitly or rather use @DirtiesContext. I do not have much experience on Kafka Streams but others can help on how to handle this.
gharchive/issue
2024-07-09T10:12:50
2025-04-01T06:45:58.954098
{ "authors": [ "MarvinVaillant", "eddumelendez" ], "repo": "testcontainers/testcontainers-java", "url": "https://github.com/testcontainers/testcontainers-java/issues/8867", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1588646833
Add windows build. Adding a windows build as in #309 is not currently possible because windows-latest does not support running linux-based images (cf. https://github.com/actions/runner-images/issues/1143). The same seems to be true for CircleCI (see https://circleci.com/docs/using-windows/#known-issues-and-limitations) and Travis (see https://travis-ci.community/t/hyper-v-support-in-windows/2605/3). Once it is feasible, running automated tests could help address #303, #108, #101, and #99. seems like this is fixed in #345?
gharchive/issue
2023-02-17T02:09:31
2025-04-01T06:45:58.957477
{ "authors": [ "alexanderankin", "tillahoffmann" ], "repo": "testcontainers/testcontainers-python", "url": "https://github.com/testcontainers/testcontainers-python/issues/310", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2149424897
Exception while running typefuzz on the remote server Bug Description While trying to use typefuzz on Ubuntu Linux Linux mantella 5.4.0-159-generic #176-Ubuntu SMP Mon Aug 14 12:04:20 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux It returns an error, trying to iterate on argument of type "Term" Note Possibly it is due to the command, as cvc5 is not installed on the machine, and I try to provide it by giving the executable path in the command: ./bin/typefuzz "z3 model_validate=true;../cvc5/build/bin/cvc5 --check-models -m -i -q" ../opensmt/regression/QF_UF/NEQ004_size4.smt│2 Hi @wintered, thanks for the quick response. I tried to run it with QF_LIA, but stubled into a different error Any ideas what could have caused it?
gharchive/issue
2024-02-22T16:12:12
2025-04-01T06:45:59.011308
{ "authors": [ "BritikovKI" ], "repo": "testsmt/yinyang", "url": "https://github.com/testsmt/yinyang/issues/66", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2238370297
[Feature Request] feedback after setting volume If i set the volume in Home Assistant, ideally it would blink the volume indicators on the device. Right now there is no feedback at all. Also Google has the device "click" as sort of a "volume preview" when you turn it up or down. It should do that when setting the volume, regardless of manner. Due to the mutex limitation in ESPHome, the click might not be possible when listening for wake word. However, I agree it would be a nice thing to temporarily flash the volume LED configuration even when the volume is changed off-device.
gharchive/issue
2024-04-11T19:00:51
2025-04-01T06:45:59.012796
{ "authors": [ "jherby2k", "tetele" ], "repo": "tetele/onju-voice-satellite", "url": "https://github.com/tetele/onju-voice-satellite/issues/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
142066910
docs(txtnode): Update document about TxtNode/TxtAST Reflect update from "Compliances tests for TxtNode #141" This update only reflect current status. It don't change anything. Rename TxtSyntax to TxtAST https://github.com/textlint/textlint/blob/master/docs/txtnode.d.ts Split TxtNodeto ParentNode(block) and TextNode(inline) Define value property to TextNode Add Compliance tests tool https://github.com/textlint/textlint-ast-test Add note about https://github.com/wooorm/unist Add notice for Processor plugin creator @homu r+ :pushpin: Commit c05803a has been approved by azu :zap: Test exempted - status
gharchive/pull-request
2016-03-19T14:34:07
2025-04-01T06:45:59.034139
{ "authors": [ "azu", "homu" ], "repo": "textlint/textlint", "url": "https://github.com/textlint/textlint/pull/170", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
593367224
Musicalypse don't load library when starting Hello! I use Musicalypse with snap on Gallium OS. The music folder has been added in settings (it's an external hard drive : /mnt...) The first scan was OK But after reboot, Musicalypse don't load this folder. He just memorizes the last song played. How to fix it please? (and : thanks for this beautiful design) Hello, Please try installing Musicalypse through the available AppImage directly. This might fix your issue. It's what I did: I downloaded the Musicalypse-1.0.0-x64.AppImage Gave perrmissions Set the external drive as source for music Launch the first scan (the library filled with folders) But when I quit and launch again Musicalype, the library looks empty. (dommage, cette appli est classe ! :) On launch, Musicalypse scans the library directory and removes any file that has been deleted or that it doesn't have access to anymore. I guess in your case the external drive might be unavailable on application start, hence the whole library is cleared. So you should make sure the external drive is indeed available whenever you launch Musicalypse. To check this assertion maybe try to use a local folder with some music and see if the problem persists. (Merci! ;) Yes, that's it: Musicalypse remembers the local music folder. Thanks! (Et continuez : cette appli est vraiment très bien !)
gharchive/issue
2020-04-03T13:03:34
2025-04-01T06:45:59.121073
{ "authors": [ "StephaneDeVannes", "tgambet" ], "repo": "tgambet/musicalypse", "url": "https://github.com/tgambet/musicalypse/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
121507485
Why cant I save a related collection? return new respond.Responses([{ response: 'response one.', created_by: 'TestUser' }, { response: 'response two.', created_by: 'TestUser' }, { response: 'response three.', created_by: 'TestUser' }]).invokeThen('save') .then(function(responses) { // This works, one at a time though // return responses[0].related('triggers').create({ // trigger: 'trigger one', // chance: 0.03, // created_by: 'TestUser2' // }); // This doesnt work // ERROR // Unhandled rejection Error: SQLITE_CONSTRAINT: NOT NULL constraint failed: trigger.response_id responses[0].related('triggers').add([{ trigger: 'trigger one', chance: 0.03, created_by: 'TestUser2' }, { trigger: 'trigger two', chance: 0.08, created_by: 'TestUser2' }]).invokeThen('save'); return; }) var bookshelf = require('bookshelf')(knexContext); var Response = bookshelf.Model.extend({ tableName: 'response', hasTimestamps: true, triggers: function() { return this.hasMany(Trigger); } }); var Trigger = bookshelf.Model.extend({ tableName: 'trigger', hasTimestamps: true, respond: function() { return this.belongsTo(Response); } }) var Responses = bookshelf.Collection.extend({ model: Response }); var Triggers = bookshelf.Collection.extend({ model: Trigger }); Setting response_id explicitly works for me: var dbFile = './test.sqlite'; var knex = require('knex')({ client: 'sqlite3', connection: { filename: dbFile } }); var bookshelf = require('bookshelf')(knex); var fs = require('fs'); if (fs.existsSync(dbFile)) { fs.unlinkSync(dbFile); } var Response = bookshelf.Model.extend({ tableName: 'response', hasTimestamps: true, triggers: function() { return this.hasMany(Trigger); } }); var Trigger = bookshelf.Model.extend({ tableName: 'trigger', hasTimestamps: true, respond: function() { return this.belongsTo(Response); } }); var Responses = bookshelf.Collection.extend({ model: Response }); var Triggers = bookshelf.Collection.extend({ model: Trigger }); return knex.schema.createTable('response', function (table) { table.increments(); table.string('response'); table.string('created_by'); table.timestamps(); }) .then(function () { return knex.schema.createTable('trigger', function (table) { table.increments(); table.integer('response_id').notNullable(); table.string('trigger'); table.float('chance'); table.string('created_by'); table.timestamps(); }); }) .then(function () { return new Responses([{ response: 'response one.', created_by: 'TestUser' }, { response: 'response two.', created_by: 'TestUser' }, { response: 'response three.', created_by: 'TestUser' }]) .invokeThen('save') .then(function(responses) { return responses[0].related('triggers').add([{ response_id: responses[0].id, trigger: 'trigger one', chance: 0.03, created_by: 'TestUser2' }, { response_id: responses[0].id, trigger: 'trigger two', chance: 0.08, created_by: 'TestUser2' }]) .invokeThen('save'); }); }) .then(function () { return Response.fetchAll({ withRelated: 'triggers' }) .then(function (responses) { console.log(require('util').inspect(responses.toJSON(), { depth: null })); }); }) .finally(function () { return knex.destroy(); }); $ node index.js [ { id: 1, response: 'response one.', created_by: 'TestUser', created_at: 1449775827903, updated_at: 1449775827903, triggers: [ { id: 1, response_id: 1, trigger: 'trigger one', chance: 0.03, created_by: 'TestUser2', created_at: 1449775827925, updated_at: 1449775827925 }, { id: 2, response_id: 1, trigger: 'trigger two', chance: 0.08, created_by: 'TestUser2', created_at: 1449775827925, updated_at: 1449775827925 } ] }, { id: 2, response: 'response two.', created_by: 'TestUser', created_at: 1449775827904, updated_at: 1449775827904, triggers: [] }, { id: 3, response: 'response three.', created_by: 'TestUser', created_at: 1449775827904, updated_at: 1449775827904, triggers: [] } ] So why does create just figure
gharchive/issue
2015-12-10T15:11:51
2025-04-01T06:45:59.124230
{ "authors": [ "LordWingZero", "blah238" ], "repo": "tgriesser/bookshelf", "url": "https://github.com/tgriesser/bookshelf/issues/1060", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2499650370
🛑 DDoS-Relay US-E-nyc1-01 is down In 5aa2c7a, DDoS-Relay US-E-nyc1-01 (http://178.128.157.126:42069/;csv) was down: HTTP code: 0 Response time: 0 ms Resolved: DDoS-Relay US-E-nyc1-01 is back up in 6b2dc92 after 5 minutes.
gharchive/issue
2024-09-01T19:42:19
2025-04-01T06:45:59.127552
{ "authors": [ "tgstation-server" ], "repo": "tgstation-operations/status", "url": "https://github.com/tgstation-operations/status/issues/582", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
996854505
Replace/revamp master branch [x] make npm the default branch [ ] rename npm to (e.g.) release [ ] check in with @layik re. unpublished commits on master [ ] delete master This is mainly to keep using GH functionality such as automatic issue closing etc. The npm development is the way forward. ~However, the original fork is useful for people who might want to use it the way the original work was done.~ Now is the time for a new master/main branch. Imported from #96: For clarification, the branch was the main dev repo for eAtlas/TGVE and has now grown out of its purpose. master branch is now way behind branch npm branch npm is the most important part, npm-dev is where I do work and create PRs on npm. Wiki is the main docs for "the project" Dockerfile is based on "old" approach of building the front-end using JS tools (npm) Must be done: README to serve as pointer to Wiki/docs and other repo docs etc master branch => npm branch? etc Consider doing: ?? #92 Closing #92 as a duplicate of this. Importing comment from that issue by @layik: Thanks @LydiaFrance and @rolyp. One thing that is prevented/preventing me from making my mind up is tgve/eatlas-template and tgve/full-template in terms of what a master branch should look like. Say we rename this repo as tgvejs (i.e this repository becomes the "source of truth"), then we would either keep those repositories or move them into an "example" folder in the same tgvejs repo and possibly build those as the GH pages output. If we do keep them, and so far I am inclined towards this, as they are now, then maybe I need to just propose to rename master as master-old and rename npm as master. The importance of those two repositories can be documented, if not already, with their reasons. Any thoughts on this would be great for me. Thanks @layik Ah, there seem to be commits on master that haven't yet been published to npm, so things aren't straightforward. In particular README.md is quite different on the two branches. Need to discuss. Indeed! @layik Btw, I will (occasionally) work weird hours, e.g. at weekends. Please ignore any comments that come your way during those unusual hours (unless you wish to respond, of course)! That makes two of us, glad you gave the warning first. @layik I'll reopen this until we're able to delete old. OK, I have pushed "old" = "master" to private repo for backup only.
gharchive/issue
2021-09-15T09:22:23
2025-04-01T06:45:59.330129
{ "authors": [ "layik", "rolyp" ], "repo": "tgve/eAtlas", "url": "https://github.com/tgve/eAtlas/issues/64", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1289673049
🛑 SoundHound is down In 30fc83c, SoundHound (https://www.soundhound.com) was down: HTTP code: 403 Response time: 97 ms Resolved: SoundHound is back up in c5487a9.
gharchive/issue
2022-06-30T06:43:06
2025-04-01T06:45:59.339440
{ "authors": [ "thaaai" ], "repo": "thaaai/upptime", "url": "https://github.com/thaaai/upptime/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
168414342
Answer what Tomi was doing with the reserved value for the change counter I've decided to check in 720 since it is passing tests. But I still need you to get an answer to this question. change counter is used to determine if we should update peer properties. if old change counter!= new change counter or old peer properties==null android bluetooth lib will update peer properties. org.thaliproject.p2p.btconnectorlib.utils.PeerModel#addOrUpdateDiscoveredPeer . So if it's new peer properties (with any value of change counter) it will be updated. Or if old change counter == 255 and new change counter ==0 .
gharchive/issue
2016-07-29T21:06:06
2025-04-01T06:45:59.360938
{ "authors": [ "vasilevskayaem", "yaronyg" ], "repo": "thaliproject/Thali_CordovaPlugin", "url": "https://github.com/thaliproject/Thali_CordovaPlugin/issues/810", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2280253635
[Bug] Non-proficient with shields not being triggered Shields should have the same disadvantages as armor when it comes to non-proficiency. Armor Proficiency. Anyone can put on a suit of armor or strap a shield to an arm. Only those proficient in the armor's use know how to wear it effectively, however. Your class gives you proficiency with certain types of armor. If you wear armor that you lack proficiency with, you have disadvantage on any ability check, saving throw, or attack roll that involves Strength or Dexterity, and you can't cast spells. Tested with different shields, and armor. All armor worked as intended, no shields cause the disadvantage. Shield Proficiency is obtained from Class Features, or the Moderately Armored feat. Foundry - 11.315 DnD5e - 3.1.2 Good point. I haven't added Shields in the mix to be honest, cause I didn't think about it. I will put it in the next release. Thanks for that.
gharchive/issue
2024-05-06T07:40:44
2025-04-01T06:45:59.381527
{ "authors": [ "glitchies", "thatlonelybugbear" ], "repo": "thatlonelybugbear/automated-conditions-5e", "url": "https://github.com/thatlonelybugbear/automated-conditions-5e/issues/125", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
668423806
Wrong compat bounds Please don't have compat bounds like https://github.com/thautwarm/NameResolution.jl/blob/e0d52d57f6487faba685517ee88ff9fbbfcc707d/Project.toml#L13. You just broke my package which is three links down the depency chain with https://github.com/thautwarm/PrettyPrint.jl/releases/tag/v0.2.0. With that release, it's no longer possible to load this package and therefore also not possible to load any of the 14 packages that depend on this package. Ref https://github.com/JuliaRegistries/General/pull/18713
gharchive/issue
2020-07-30T06:47:31
2025-04-01T06:45:59.383538
{ "authors": [ "andreasnoack" ], "repo": "thautwarm/NameResolution.jl", "url": "https://github.com/thautwarm/NameResolution.jl/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
475160614
ReleaseNewVersion does not work as a user command. Fixed
gharchive/issue
2019-07-31T13:56:20
2025-04-01T06:45:59.391017
{ "authors": [ "PaulMansour" ], "repo": "the-carlisle-group/AcreTools", "url": "https://github.com/the-carlisle-group/AcreTools/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
253723900
Question? The next release will be for Laravel 5.5, or for both versions 5.4 & 5.5? And please dont forget that many dependencies doesn't support Laravel 5.5(If you suggest migrating from 5.4 to 5.5) for now and dont forget your old users of Voyager to have new cool stuf The new version of Voyager (1.0) will support Laravel 5.3, 5.4, and 5.5. It will not require any specific version to run, so it will not affect your project or its other dependencies. Super @fletch3555 why did not I write earlier)))))))
gharchive/issue
2017-08-29T16:18:20
2025-04-01T06:45:59.396704
{ "authors": [ "davodavodavo3", "fletch3555" ], "repo": "the-control-group/voyager", "url": "https://github.com/the-control-group/voyager/issues/1614", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
256201972
Permissions and menu items Laravel Version: 5.5 Voyager Version: 1.0 dev PHP Version: 7.1.1 Database Driver & Version: MySQL 5.6.12 Description: Some items in the sidebar menu are not hidden, as they should according to the permissions I defined in the Roles page. I noticed this problem for the Tools dropdown (if empty, it is visible but empty) and the Settings and Hooks pages. Steps To Reproduce: From a fresh installation, go to roles and edit permissions for any of them: uncheck Browse Database, Browse Compass, Menus, Settings and Admin/Browse Hooks permissions. After signing in with a user having this edited role, the Tools dropdown, Settings and Hooks items are still visible. Laravel Version: 5.5 Voyager Version: 1.0 dev PHP Version: 7.1.5 Database Driver & Version: MySQL 5.7.17 I have same problem. @fletch3555 , @jeremyargoud Looks like I've fixed the problem. But the code that checks the permissions must refactor to a fuction in package. I dont know where can i put it. So i dont send pull request. File: vendor/tcg/voyager/resources/views/menu/admin_menu.blade.php ` @php if (Voyager::translatable($items)) { $items = $items->load('translations'); } @endphp @foreach ($items as $item) @php $originalItem = $item; if (Voyager::translatable($item)) { $item = $item->translate($options->locale); } // TODO - still a bit ugly - can move some of this stuff off to a helper in the future. $listItemClass = []; $styles = null; $linkAttributes = null; if(url($item->link()) == url()->current()) { array_push($listItemClass,'active'); } //If have children, default false. we will check childs permissions! if(!$originalItem->children->isEmpty()) { $show_menu_item = false; } else { $show_menu_item = true; } // With Children Attributes if(!$originalItem->children->isEmpty()) { foreach($originalItem->children as $children) { if(url($children->link()) == url()->current()) { array_push($listItemClass,'active'); } // Childrens Permission Checker $self_prefix = str_replace('/', '\/', $options->user->prefix); $slug = str_replace('/', '', preg_replace('/^\/'.$self_prefix.'/', '', $children->link())); if ($slug != '') { // Get dataType using slug $dataType = $options->user->dataTypes->first(function ($value) use ($slug) { return $value->slug == $slug; }); if ($dataType) { // Check if datatype permission exist $exist = $options->user->permissions->first(function ($value) use ($dataType) { return $value->key == 'browse_'.$dataType->name; }); } else { // Check if admin permission exists $exist = $options->user->permissions->first(function ($value) use ($slug) { return $value->key == 'browse_'.$slug; }); } if ($exist) { if (in_array($exist->key, $options->user->user_permissions)) { $show_menu_item = true; } } } } $linkAttributes = 'href="#' . str_slug($item->title, '-') .'-dropdown-element" data-toggle="collapse" aria-expanded="'. (in_array('active', $listItemClass) ? 'true' : 'false').'"'; array_push($listItemClass, 'dropdown'); } else { $linkAttributes = 'href="' . url($item->link()) .'"'; } // Permission Checker $self_prefix = str_replace('/', '\/', $options->user->prefix); $slug = str_replace('/', '', preg_replace('/^\/'.$self_prefix.'/', '', $item->link())); if ($slug != '') { // Get dataType using slug $dataType = $options->user->dataTypes->first(function ($value) use ($slug) { return $value->slug == $slug; }); if ($dataType) { // Check if datatype permission exist $exist = $options->user->permissions->first(function ($value) use ($dataType) { return $value->key == 'browse_'.$dataType->name; }); } else { // Check if admin permission exists $exist = $options->user->permissions->first(function ($value) use ($slug) { return $value->key == 'browse_'.$slug; }); } if ($exist) { // Check if current user has access if (!in_array($exist->key, $options->user->user_permissions)) { continue; } } } @endphp @if($show_menu_item) <li class="{{ implode(" ", $listItemClass) }}"> <a {!! $linkAttributes !!} target="{{ $item->target }}"> <span class="icon {{ $item->icon_class }}"></span> <span class="title">{{ $item->title }}</span> </a> @if(!$originalItem->children->isEmpty()) <div id="{{ str_slug($originalItem->title, '-') }}-dropdown-element" class="panel-collapse collapse {{ (in_array('active', $listItemClass) ? 'in' : '') }}"> <div class="panel-body"> @include('voyager::menu.admin_menu', ['items' => $originalItem->children, 'options' => $options, 'innerLoop' => true]) </div> </div> @endif </li> @endif @endforeach ` No need to post the whole file.... Please only post the suggested changes, or better yet, open a PR with the changes for us to review @fletch3555 Thank you for fast feed back. PR for you to review. https://github.com/the-control-group/voyager/pull/1783
gharchive/issue
2017-09-08T09:56:28
2025-04-01T06:45:59.405879
{ "authors": [ "evrend", "fletch3555", "jeremyargoud" ], "repo": "the-control-group/voyager", "url": "https://github.com/the-control-group/voyager/issues/1684", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
285046603
MacOS new features. translate English to other languages save notes create files Great going! Fix the typos in activities.md If I am not wrong Mac has Pages,Keynote and Numbers for the Microsoft equivalents, right? yeah ! sorry , dint check those . Great!
gharchive/pull-request
2017-12-29T07:23:27
2025-04-01T06:45:59.408398
{ "authors": [ "Snehal-Reddy", "the-ethan-hunt" ], "repo": "the-ethan-hunt/B.E.N.J.I.", "url": "https://github.com/the-ethan-hunt/B.E.N.J.I./pull/96", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1599689462
Add translation of PIDINST schema to pydantic This PR simply adds a pydantic model for instruments that is a direct translation of the RDA PIDINST schema https://github.com/rdawg-pidinst/schema/blob/master/schema.rst. They themselves provide two versions that includes a translation to DataCite's format. It's not clear how they expect this data to be reused. I don't like the resource layout (it is very XML-y) so think we should use this as an example for having our own internally schema but being able to map out to others. I may offer to contribute this file back upstream to PIDINST (maybe as LinkML yaml instead) and see what they say. Superseded by #531
gharchive/pull-request
2023-02-25T12:10:12
2025-04-01T06:45:59.412455
{ "authors": [ "ml-evs" ], "repo": "the-grey-group/datalab", "url": "https://github.com/the-grey-group/datalab/pull/304", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2054770436
fix price of gpu shown on btc farm cost page Used to show the GPU sell price, now properly shows the GPU buy price. .deploy .deploy .deploy .deploy .deploy .deploy
gharchive/pull-request
2023-12-23T12:14:17
2025-04-01T06:45:59.414177
{ "authors": [ "Razzmatazzz" ], "repo": "the-hideout/tarkov-dev", "url": "https://github.com/the-hideout/tarkov-dev/pull/792", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2368226298
1.21 Update No rush, I know the version just came out. Just a nudge. It's been a godsend on my server in 1.20.6 and I'd be honored to have it in the server for 1.21 :) Done. Done.
gharchive/issue
2024-06-23T04:27:12
2025-04-01T06:45:59.430394
{ "authors": [ "Proxi2", "the-sh4d0w" ], "repo": "the-sh4d0w/too-cheap", "url": "https://github.com/the-sh4d0w/too-cheap/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2628096289
🛑 glance-probe is down In 14512a4, glance-probe (https://glance.api.sjc3.rackspacecloud.com) was down: HTTP code: 502 Response time: 9511 ms Resolved: glance-probe is back up in b22cf93 after 1 hour, 33 minutes.
gharchive/issue
2024-11-01T01:06:02
2025-04-01T06:45:59.439012
{ "authors": [ "the2hill" ], "repo": "the2hill/rs-flex-uptime", "url": "https://github.com/the2hill/rs-flex-uptime/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
206084194
Add Database Setup Changes Add docker postgres start and stop scripts Add database migrations Please Review: @sagersmith8 +1 looks pretty good, could we maybe put them in their own folder though? Design Database
gharchive/pull-request
2017-02-08T03:15:37
2025-04-01T06:45:59.449129
{ "authors": [ "Kuroshii", "sagersmith8" ], "repo": "theVite/vite_server", "url": "https://github.com/theVite/vite_server/pull/18", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
119724448
Use with Apache Phoenix I would like to output events from logstash to HBase, via Apache Phoenix: https://phoenix.apache.org My output configuration: jdbc { connection_string => 'jdbc:phoenix://hbase' statement => [ "upsert into test values (2,'World!')"] } I can see JAR is loaded because duplicate SLF4J errors: SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.1/vendor/jar-dependencies/runtime-jars/slf4j-nop-1.7.13.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/opt/logstash/vendor/jar/jdbc/phoenix-4.6.0-HBase-1.1-client.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.helpers.NOPLoggerFactory] Error: { : outgoing_count=>1, : exception=>"Java::JavaLang::RuntimeException", : backtrace=>[ "com.zaxxer.hikari.util.DriverDataSource.<init>(com/zaxxer/hikari/util/DriverDataSource.java:88)", "com.zaxxer.hikari.pool.PoolBase.initializeDataSource(com/zaxxer/hikari/pool/PoolBase.java:296)", "com.zaxxer.hikari.pool.PoolBase.<init>(com/zaxxer/hikari/pool/PoolBase.java:84)", "com.zaxxer.hikari.pool.HikariPool.<init>(com/zaxxer/hikari/pool/HikariPool.java:103)", "com.zaxxer.hikari.HikariDataSource.getConnection(com/zaxxer/hikari/HikariDataSource.java:94)", "java.lang.reflect.Method.invoke(java/lang/reflect/Method.java:497)", "LogStash::Outputs::Jdbc.safe_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.1/lib/logstash/outputs/jdbc.rb:168)", "LogStash::Outputs::Jdbc.safe_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.1/lib/logstash/outputs/jdbc.rb:168)", "LogStash::Outputs::Jdbc.flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.1/lib/logstash/outputs/jdbc.rb:116)", "LogStash::Outputs::Jdbc.flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.1/lib/logstash/outputs/jdbc.rb:116)", "Stud::Buffer.buffer_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:219)", "Stud::Buffer.buffer_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:219)", "org.jruby.RubyHash.each(org/jruby/RubyHash.java:1342)", "Stud::Buffer.buffer_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:216)", "Stud::Buffer.buffer_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:216)", "Stud::Buffer.buffer_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:193)", "Stud::Buffer.buffer_flush(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:193)", "RUBY.buffer_receive(/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:159)", "RUBY.receive(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.1/lib/logstash/outputs/jdbc.rb:109)", "RUBY.handle(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.1.0-java/lib/logstash/outputs/base.rb:81)", "RUBY.output_func((eval):323)", "RUBY.outputworker(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.1.0-java/lib/logstash/pipeline.rb:276)", "RUBY.start_outputs(/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.1.0-java/lib/logstash/pipeline.rb:193)", "java.lang.Thread.run(java/lang/Thread.java:745)" ], : level=>: warn } Any idea? Thanks, The SLF4J warning isn't ideal, but I don't think it's should actually be causing a problem. That said, I can't stand up a hbase instance to check today though :disappointed: Couple of questions: Does the error happen when logstash starts, or when an event is tried to be sent? Looking at https://phoenix.apache.org/ the example connection strings don't include the preceding slashes before the hostname. I won't pretend to have ever looked at phoenix before - but have you tried without them? I don't know if phoenix is perhaps touchy about them? Looks like connection string jdbc:phoenix:hbase:/hbase works actually ! I am not getting errors any more. I keep you posted if it's working, that could mean you wrote a very nice plugin to output to HBase! I'd love to know if it does or doesn't work :+1: If you have problems, let me know and I'll see what I can do. If it does work if you could put together a sample configuration that would be amazing :smile: Haaa https://github.com/theangryangel/logstash-output-jdbc/blob/master/lib/logstash/outputs/jdbc.rb#L173 Took me some hours to figure out this! So with statements it's working fine! data are inserted into HBase, cool! Thanks, Ah sorry about that @ebuildy. I clearly didn't look at your example configuration very well :( The code does assume that you're actually passing in some event data, which is why it has that check. Would it have been helpful if we output a warning before skipping the event? Or do you have a use case where you wouldnt actually insert any event data (other than testing)? :) Nop, was just for testing. But a warning could save some lives in the futur ;-) Added in 529c98aadb91dcb6e8a3db8ae976ad543178e8b8
gharchive/issue
2015-12-01T14:09:52
2025-04-01T06:45:59.509830
{ "authors": [ "ebuildy", "theangryangel" ], "repo": "theangryangel/logstash-output-jdbc", "url": "https://github.com/theangryangel/logstash-output-jdbc/issues/22", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
937191589
typescript d.ts 和导入误报错 1.克隆任何一个 typescript 源码库,文件中的导入都会报错找不到模块。 2.d.ts 中 declare namespace cc 会报错 duplicate identifier cc 以上两个问题在 vscode 是正常的 3.应用里显示的是3.7.4,是否只是显示错误 4.下次大版本更新是否能更新到typescript 4.3,增加了 override ,很有用 感谢这么快速地更新到了4.2 这个有关多文件的支持,需要透过未来更新支援。目前编辑器只会分析已经开启的档案。 应该是bug,会尽快修复 对的,只是显示错误。编辑器的TypeScript版本需要透过app更新,但terminal内的tsc 可以用npm install -g typescript自行更新 4. 会的 感谢,比较致命的问题是第一点,导致现在我就算只是想在 ipad 上用 codeapp 预览一下代码都不太实用,希望能解决 明白,会尽量改善这问题。 能否进一步更新ts版本,我的项目使用了moduleResolution: NodeNext,目前完全不能正常使用。 我对 ios 开发很不熟悉,所以我没找到 TS 的引入位置,希望能够尽快得到支持(顺便改进了一下中文本地化) 我不认为这个问题被修复了,如果你觉得你已经不存在原问题,计划关闭的话,那我就新开一个。 @bummoblizard 感谢,请问有计划更新一下内置的ts版本么?不太清楚有什么困难,但至少ts不应该有ios适配方面的问题。 能够在ipad上临时应急改个项目还是不错的,(虽然已经有vscode.dev了) @Mister-Hope 上游 Monaco Editor 仍在4.5.5。如果你真的需要,我可以手动更新它。 https://github.com/microsoft/monaco-editor/blob/f20e0c8d472bababab79108dff951ae09460c4ed/src/language/typescript/lib/typescriptServicesMetadata.ts#L5 恩恩,还是建议更新一下吧😅因为我在使用一些新版本的特性。 恩恩,还是建议更新一下吧😅因为我在使用一些新版本的特性。 懒得跟进了,作者有空就加吧
gharchive/issue
2021-07-05T15:21:32
2025-04-01T06:45:59.552865
{ "authors": [ "Mister-Hope", "TheSmallMain", "bummoblizard", "smallmain" ], "repo": "thebaselab/codeapp", "url": "https://github.com/thebaselab/codeapp/issues/179", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1025031378
Impossible to track events on iOS 🐛 Bug Report I'm trying to log events from my app, but the events are not shown correctly in the events manager dashboard of Facebook. I'm using all the settings from the Setup / Installation, i'm using the same FB account on the test device ( iPhone XR, iOS 15 ) and the only events that i can see are "fb_mobile_login_method_start" and "fb_mobile_login_start". To Reproduce / Code Example This is the javascript function that i'm calling for test. After loading the main screen Settings.initializeSDK(); await Settings.setAdvertiserTrackingEnabled(true); By pressing one button AppEventsLogger.logPurchase(Math.random(999), 'EUR', { [AppEventsLogger.AppEventParams.ContentID]: `order-${Math.random( 9999, ).toString()}`, }); Expected Behavior The strange thing is that with the account of a colleague , the events are being tracked. Environment OS: macOS 11.6 CPU: (8) x64 Apple M1 Memory: 20.29 MB / 16.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 15.6.0 - ~/.nvm/versions/node/v15.6.0/bin/node Yarn: 1.22.15 - /usr/local/bin/yarn npm: 7.4.0 - ~/.nvm/versions/node/v15.6.0/bin/npm Watchman: 2021.10.04.00 - /usr/local/bin/watchman Managers: CocoaPods: 1.11.2 - /usr/local/bin/pod SDKs: iOS SDK: Platforms: iOS 15.0, DriverKit 20.4, macOS 11.3, tvOS 15.0, watchOS 8.0 Android SDK: Not Found IDEs: Android Studio: 2020.3 AI-203.7717.56.2031.7678000 Xcode: 13.0/13A233 - /usr/bin/xcodebuild Languages: Java: Not Found npmPackages: @react-native-community/cli: Not Found react: 17.0.2 => 17.0.2 react-native: 0.65.1 => 0.65.1 react-native-macos: Not Found npmGlobalPackages: *react-native*: Not Found This is not actionable without some sort of reproducible example, apologies https://stackoverflow.com/help/minimal-reproducible-example We are about to release with a bump of underlying ios sdk from 9 to 11, maybe that helps if not we must have an example that reproducibly fails for anyone to offer any help
gharchive/issue
2021-10-13T09:50:40
2025-04-01T06:45:59.559086
{ "authors": [ "Derewith", "mikehardy" ], "repo": "thebergamo/react-native-fbsdk-next", "url": "https://github.com/thebergamo/react-native-fbsdk-next/issues/114", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2614138646
Minor error handing and tracing improvements Rename update to get_line_value Add specific exceptions when hub is not online and user is using the switches Remove the duplicate update_lines call in the cover case @thecode can you take a look again? I added another submit with README changes.
gharchive/pull-request
2024-10-25T13:38:39
2025-04-01T06:45:59.578792
{ "authors": [ "tomer-w" ], "repo": "thecode/ha-rpi_gpio", "url": "https://github.com/thecode/ha-rpi_gpio/pull/308", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
347265678
Docker swarm machine I have created a Swarm and consule. After that i tried to do below comment but i am having doubt that what is a and b in the comment mentioned. "Now we can configure our new Swarm agents on each host a and b: $ docker run -d --restart=always --name swarm-agent-consul swarm:1.0.0 join --advertise $(curl http://169.254.169.254/latest/meta-data/public-ipv4):2376 consul://192.169.0.1:8500" Also when i used this command i am getting error as mentioned below: Hi @Ajith-AK! Thank you for reaching out, there are a few things to note here. This repository is not being maintained as The {code} Team dissolved in March 2018, see https://blog.thecodeteam.com/2018/02/22/final-thank-code-team/. I believe you meant to create an issue, not issue a pull request. The pull request includes a great number of changes from @nikolaia and @mastoj from years ago, unfortunately none of which has been sent back to the main repository (I wish they had been, the changes look really great). This workshop is ~3 years old, so things will most likely not work exactly as they did during that time. Final thought, your command includes an IP address that's not private, 192.169.0.1, you should be using a private block such as 192.168.0.0 - 192.168.255.255 when seeing up your environment. Best of luck! Thanks
gharchive/pull-request
2018-08-03T05:39:04
2025-04-01T06:45:59.583668
{ "authors": [ "Ajith-AK", "jonasrosland" ], "repo": "thecodeteam/training", "url": "https://github.com/thecodeteam/training/pull/60", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
433838714
Adding Deferred types support We should add support for Webonyx Deferred types in order to offer a way to avoid the N+1 issues. Implementation should be as easy as possible on the user side. Current proposal: Consider a request where a user attached to a post must be returned. { posts { id user { id } } } /** * @Type */ class PostType { /** * @Field(prefetchMethod="prefetchUsers") * @param Post $post * @param mixed $prefetchedUsers * @return User */ public function getUser(Post $post, $prefetchedUsers): User { // This method will receive the $prefetchedUsers as second argument. This is the return value of the "prefetchUsers" method below. // Using this prefetched list, it should be easy to map it to the post } /** * @param Post[] $posts * @return mixed */ public function prefetchUsers(iterable $posts) { // This function is called only once per GraphQL request // with the list of posts. You can fetch the list of users // associated with this posts in a single request, // for instance using a "IN" query in SQL or a multi-fetch // in your cache back-end. } } When the "prefetchMethod" attribute is detected in the "@Field" annotation, the method is called automatically. The first argument of the method is an iterable over the instances of the main type. The "prefetchMethod" can return absolutely anything (mixed). The return value will be passed as the second parameter of the "@Field" annotated method. Important: field arguments can be set either on the @Field annotated method OR/AND on the prefetchMethod. For instance: /** * @Type */ class PostType { /** * @Field(prefetchMethod="prefetchComments") * @param Post $post * @param mixed $prefetchedComments * @return Comment[] */ public function getComments(Post $post, $prefetchedComments): array { // ... } /** * @param Post[] $posts * @return mixed */ public function prefetchComments(iterable $posts, bool $hideSpam, int $filterByScore) { // Parameters passed after the first parameter (hideSpam, filterByScore...) are automatically exposed as GraphQL arguments for the "comments" field. } } The prefetch method MUST be in the same class as the @Field-annotated method and MUST be public. Note: the "prefetchMethod" argument CANNOT be applied on @Query and @Mutation annotations. We will need a different kind of attribute for this. Note to self: implementation warning! A GraphQL query can contain many different calls with different arguments for a same field. For instance: { posts { id legitComments: comments(hideSpam: true) { id } spamComments: comments(hideSpam: false) { id } } } With the example above, we will have to call the prefetch method twice! Once with the "hideSpam" set to true, and once with the "hideSpam" set to false. To be honest, this format seems a little bit off. What happened with the original where getComments(Post $post): array would actually be getComments(array $posts): array? Maybe I just don't see exactly how does it work, but let me try though: <?php /** * @Type */ class PostType { /** * @Field(prefetchMethod="eagerLoadComments") * @return Comment[] */ public function getComments(Post $post, $promisedData): array { return array_filter($promisedData, function ($comment) use ($post) { return $post->getId() === $comment['post_id']; }); } /** * @param Post[] $posts * @return array */ public function eagerLoadComments(iterable $posts, bool $hideSpam, int $filterByScore) { return [ [ 'post_id' => 1, 'comment' => 'Hello world', ], [ 'post_id' => 1, 'comment' => 'Hello world2', ], [ 'post_id' => 2, 'comment' => 'Hello world3', ], ]; } } Is this what you wanted to achieve? Hey @golaod , Your example is perfect, you got it right. Indeed, the idea we first talked about was different. My first idea was something like: /** * @Type */ class PostType { /** * @BatchField() * @param Post[] $posts * @return Comment[][] */ public function getComments(array $posts): array { // ... } } I've been talking with @homersimpsons for quite some time about it. He convinced me this is a somewhat better alternative. Here are the main points: Regarding method signature /** * @Field(prefetchMethod="eagerLoadComments") * @return Comment[] */ public function getComments(Post $post, $promisedData): array This method signature is quite similar to the usual way we create fields. It is easier to understand (it is obvious that the field "comments" is of type "Comment[]". Compare this to: /** * @BatchField() * @param Post[] $posts * @return Comment[][] */ public function getComments(array $posts): array { // ... } The return type ("Comment[][]") will map to Comment[] in GraphQL. It is obvious to both of us, but it could cause problems to beginners. Regarding implementation With this signature: public function getComments(array $posts): array the keys of the returned array must match the keys of the $posts array (otherwise, GraphQLite cannot know by itself what comments are linked to what posts). Doing the matching on the keys of the arrays is error-prone. Also, what is GraphQLite supposed to return if some keys are missing or if there are some additional keys passed by the developer? Should it trigger an error? Compare this to the solution proposed by @homersimpsons: /** * @Field(prefetchMethod="eagerLoadComments") * @return Comment[] */ public function getComments(Post $post, $promisedData): array Here, there is absolutely no possible doubt. The returned comments are matching the "$post" passed in argument. The method signature is way more strict than in the other solution and less things can go awkward. Does it make sense? Or do you think another way of doing things we did not think about? How does it now about $promisedData? Asking because the field can still have inputs /** * @Field(prefetchMethod="eagerLoadComments") * @return Comment[] */ public function getComments(Post $post, $promisedData): array GraphQLite will detect the prefetchMethod attribute in the @Field annotation. If the prefetchMethod attribute is present, the 2nd argument of the method MUST be the promisedData (returned by the prefetch method). Input arguments come either in third position and afterwards, or in the prefetch method: /** * @Type */ class PostType { /** * @Field(prefetchMethod="prefetchComments") * @param Post $post * @param mixed $prefetchedComments * @return Comment[] */ public function getComments(Post $post, $prefetchedComments, int $maxCommentLength): array { // In this sample, the input arguments of the comments field are: // - maxCommentLength: Int // - hideSpam: Bool // - filterByScore: Int } /** * @param Post[] $posts * @return mixed */ public function prefetchComments(iterable $posts, bool $hideSpam, int $filterByScore) { // Parameters passed after the first parameter (hideSpam, filterByScore...) are automatically exposed as GraphQL arguments for the "comments" field. } } :+1:
gharchive/issue
2019-04-16T15:16:49
2025-04-01T06:45:59.598772
{ "authors": [ "golaod", "moufmouf" ], "repo": "thecodingmachine/graphqlite", "url": "https://github.com/thecodingmachine/graphqlite/issues/42", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
180847627
Change licence text in README to match LICENSE/metadata The commit history shows the module has always been under the ASL-2.0, and that the README text was an inconsistency in 999eb91 where the LICENSE document was added and README.md was extended (probably from a template). Updated the main LICENSE document to include the full text, not only the boilerplate copyright statement. Fixes GH-66 Thanks!
gharchive/pull-request
2016-10-04T09:40:53
2025-04-01T06:45:59.714399
{ "authors": [ "domcleal", "ekohl" ], "repo": "theforeman/puppet-dns", "url": "https://github.com/theforeman/puppet-dns/pull/69", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
372293068
AbstractDataSource methods access level Methods in AbstractDataSource are not open so in case we need to change / add something to current implementation, we can't. I wanted to ask what's the reason for such decision? Can I make PR with open methods? I agree this could be required. Do you have the reason/use-case for it so I am sure this is required? I needed to call method on table view's instance every time cellForRowAtIndexPath: is called. However, I realized it's not good at all and found different solution. I decided to close this issue because I can't think of any other scenario when we might need these public methods to be open. Thank you I'm sorry for late reply. I haven't received notification. OK, thanks for the information.
gharchive/issue
2018-10-21T08:50:17
2025-04-01T06:45:59.730407
{ "authors": [ "adam-leitgeb", "mkj-is" ], "repo": "thefuntasty/CellKit", "url": "https://github.com/thefuntasty/CellKit/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1580995242
Does this Action support something like: "feat(whatever): Issue-Name (#ISSUE-IDENTIFIER)"? Does this action support this linting? For Example: "feat(authentication): Creating Login UI + Logic(#GGG-2)" or "test: Implement Unit tests for Authentication(#GG-5)" If you can figure out a regex for it, it should @thehanimo Alright. So that is my regex: (fix|feat|build|chore|ci|docs|style|refactor|test)(\(([^)]+)\))?:\s.*\s\(#\w{1,3}-\d+\) How should my json look like because I already set the prefixes at another point of the config or is that optional? prefixes is optional so you can either delete it or pass in an empty array. Your regexp field under CHECKS would look like this: "regexp": "(fix|feat|build|chore|ci|docs|style|refactor|test)(\\(([^)]+)\\))?:\\s.*\\(#\\w{1,3}-\\d+\\)" Notice how double backslashes are required to put literal \ characters in the string. Also, I noticed a bug in your regexp where you have an extra \s before (#ISSUE-IDENTIFIER) but I'm not sure if you meant that. If you did, then your examples given above wouldn't work. Alright, I'll take a look and let you know :) One more question: What is the property regexpFlags and is it optional? Is ignoreLabels also optional? I just tried it and it works perfectly. Thanks for the great tool. Yes, regexpFlags and ignoreLabels are both optional. You can find more about the flags here.
gharchive/issue
2023-02-11T20:00:57
2025-04-01T06:45:59.739734
{ "authors": [ "NicoMarksman", "thehanimo" ], "repo": "thehanimo/pr-title-checker", "url": "https://github.com/thehanimo/pr-title-checker/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
415211547
Unable to slice with non-categorical obs index import anndata import numpy as np import pandas as pd adata = anndata.AnnData(X=np.array([[1,2], [1,2], [1,2]]), obs=pd.DataFrame(index=[0,1,2]), var=['a','b']) subset = adata[:,0] 'Don’t call _normalize_index with non-categorical/string names' AssertionError: Don’t call _normalize_index with non-categorical/string names I am getting the same error when attempting to slice with the string var name as well. subset = adata[:,'a'] subset = adata[:,adata.var_names[0]] subset = adata[:,'0'] all throw the same error Hi, i will check. Thanks, Sergei! The main problem here is that AnnData is in an invalid state and we should have thrown errors in the first place. I'll take care of it. Closing this, since it's an unsupported use case. what the heck this was a use case in your official tutorial
gharchive/issue
2019-02-27T16:26:06
2025-04-01T06:45:59.761631
{ "authors": [ "Koncopd", "cryptovec", "falexwolf", "ivirshup", "joshua-gould", "snyquist2" ], "repo": "theislab/anndata", "url": "https://github.com/theislab/anndata/issues/120", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
598872828
Error in read seurat loom object in scanpy Hello, I converted a Seurat object into a .loom file and tried to read it into Scanpy using the read_loom() function. But I got this error: OSError Traceback (most recent call last) ~/anaconda3/lib/python3.7/site-packages/h5py/_hl/files.py in make_fid(name, mode, userblock_size, fapl, fcpl, swmr) 198 try: --> 199 fid = h5f.open(name, h5f.ACC_RDWR, fapl=fapl) 200 except IOError: h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/h5f.pyx in h5py.h5f.open() OSError: Unable to open file (unable to lock file, errno = 11, error message = 'Resource temporarily unavailable') During handling of the above exception, another exception occurred: OSError Traceback (most recent call last) ~/anaconda3/lib/python3.7/site-packages/h5py/_hl/files.py in make_fid(name, mode, userblock_size, fapl, fcpl, swmr) 201 try: --> 202 fid = h5f.open(name, h5f.ACC_RDONLY, fapl=fapl) 203 except IOError: h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/h5f.pyx in h5py.h5f.open() OSError: Unable to open file (unable to lock file, errno = 11, error message = 'Resource temporarily unavailable') During handling of the above exception, another exception occurred: OSError Traceback (most recent call last) <ipython-input-13-9ead8199ac42> in <module> ----> 1 adata=sc.read_loom("/home/user/Data/pseudo_time/data/5ns+5s/seurat_5ns_5s_sae.loom") ~/anaconda3/lib/python3.7/site-packages/anndata/_io/read.py in read_loom(filename, sparse, cleanup, X_name, obs_names, obsm_names, var_names, varm_names, dtype, **kwargs) 189 from loompy import connect 190 --> 191 with connect(filename, "r", **kwargs) as lc: 192 if X_name not in lc.layers.keys(): 193 X_name = "" ~/anaconda3/lib/python3.7/site-packages/loompy/loompy.py in connect(filename, mode, validate, spec_version) 1147 Note: if validation is requested, an exception is raised if validation fails. 1148 """ -> 1149 return LoomConnection(filename, mode, validate=validate, spec_version=spec_version) ~/anaconda3/lib/python3.7/site-packages/loompy/loompy.py in __init__(self, filename, mode, validate, spec_version) 82 if validate: 83 lv = loompy.LoomValidator(version=spec_version) ---> 84 if not lv.validate(filename): 85 raise ValueError("\n".join(lv.errors) + f"\n{filename} does not appead to be a valid Loom file according to Loom spec version '{spec_version}'") 86 ~/anaconda3/lib/python3.7/site-packages/loompy/loom_validator.py in validate(self, path, strictness) 46 """ 47 valid1 = True ---> 48 with h5py.File(path) as f: 49 valid1 = self.validate_spec(f) 50 if not valid1: ~/anaconda3/lib/python3.7/site-packages/h5py/_hl/files.py in __init__(self, name, mode, driver, libver, userblock_size, swmr, rdcc_nslots, rdcc_nbytes, rdcc_w0, track_order, **kwds) 406 fid = make_fid(name, mode, userblock_size, 407 fapl, fcpl=make_fcpl(track_order=track_order), --> 408 swmr=swmr) 409 410 if isinstance(libver, tuple): ~/anaconda3/lib/python3.7/site-packages/h5py/_hl/files.py in make_fid(name, mode, userblock_size, fapl, fcpl, swmr) 202 fid = h5f.open(name, h5f.ACC_RDONLY, fapl=fapl) 203 except IOError: --> 204 fid = h5f.create(name, h5f.ACC_EXCL, fapl=fapl, fcpl=fcpl) 205 else: 206 raise ValueError("Invalid mode; must be one of r, r+, w, w-, x, a") h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/h5f.pyx in h5py.h5f.create() OSError: Unable to create file (unable to open file: name = '/home/user/Data/pseudo_time/data/5ns+5s/seurat_5ns_5s_sae.loom', errno = 17, error message = 'File exists', flags = 15, o_flags = c2) It looks like that file was open from another process while you were trying to read it. Try making sure that nothing else was trying to access the file when you read it. IIRC, a common cause for this error is needing to close the r session that wrote the file. Updating your version of loompy might work too. @ivirshup Thank you for your help. By closing R, it works fine now https://satijalab.org/loomr/loomr_tutorial It looks like that file was open from another process while you were trying to read it. Try making sure that nothing else was trying to access the file when you read it. IIRC, a common cause for this error is needing to close the r session that wrote the file. If closing the R session doesn't work, I found that the "close_all()" function works. For example "loom_file$close_all()" Source: https://satijalab.org/loomr/loomr_tutorial
gharchive/issue
2020-04-13T12:53:23
2025-04-01T06:45:59.766660
{ "authors": [ "colinmcgovern", "ivirshup", "roofya" ], "repo": "theislab/anndata", "url": "https://github.com/theislab/anndata/issues/354", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
152968152
revise zh-Hans translation There is no such idiom as "昨日黄花". The correct word should be "明日黄花". See: https://github.com/thejameskyle/babel-handbook/blob/master/CONTRIBUTING.md#translating-the-handbook
gharchive/pull-request
2016-05-04T09:24:54
2025-04-01T06:45:59.812150
{ "authors": [ "ArlenKou", "thejameskyle" ], "repo": "thejameskyle/babel-handbook", "url": "https://github.com/thejameskyle/babel-handbook/pull/76", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
381648504
Class 'Kordy\Ticketit\TicketitServiceProvider' not found i have pasted Kordy\Ticketit\TicketitServiceProvider::class, ni my config/app.php but when i rust url to install its giving following error. Class 'Kordy\Ticketit\TicketitServiceProvider' not found Was the composer install successful? You could also try composer dump-autoload yes i tried composer dump-autoloadbut not worked What version of ticketit did you install? (You can find the version number in composer.lock)
gharchive/issue
2018-11-16T15:34:57
2025-04-01T06:45:59.816067
{ "authors": [ "balping", "waqarali09" ], "repo": "thekordy/ticketit", "url": "https://github.com/thekordy/ticketit/issues/504", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
186483606
demo is included in installed package The demo package gets included into the installed package (i.e. into site-packages) since the packages clause doesn't tell the tooling not to include it. I think that's not optimal :) Could you submit a merge request to have it not part of the package? Done deal, @bufke. Sorry it took me a while to get back to doing that.
gharchive/issue
2016-11-01T08:38:22
2025-04-01T06:45:59.818018
{ "authors": [ "akx", "bufke" ], "repo": "thelabnyc/wagtail_blog", "url": "https://github.com/thelabnyc/wagtail_blog/issues/33", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }