Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
+100You could try this:annotations: {{- range $key, $value := .Values.ingress.annotations }} {{ $key }}: {{ $value | quote }} {{- end }}
I am trying to quote my annotation values. I am trying like thisannotations: {{- range $key, $value := .Values.ingress.annotations }} {{ $key }}: {{ printf "%s" $value | quote }} {{- end }}and thisannotations: {{- range $key, $value := .Values.ingress.annotations }} {{ $key }}: "{{ $value }}" {{- end ...
helm double quote annotations value
That should be possible using standard git tools. Add something like this to your workflow: cd wherever_your_output_resides # clone the public repo, or pull if we have a clone already git pull origin || git clone <url> . # check out the desired branch (optional) git checkout some_branch # stage output # (you may need ...
I need to commit the output from my private repo workflow ( github-actions) to a public repo is it possible?
How to commit output from private workflow to another public repo
On your site at port 80, you just respond with a 301 response code redirecting the user to your HTTPS site at port 443. The secure site then send the "Strict-Transport-Security" header. This will still leave your users vulnerable to man-in-the-middle attacks the very first time they visit your site. You can only miti...
Currently in the process of setting up a new personal server. I've been reading about HSTS (thanks EFF!), as well as the steps for implementing on Nginx (ex: here). What I haven't seen clearly spelled out is how to handle the initial redirect. Do I serve some static error content at port 80, redirecting to the actual...
How to Handle Port Redirection for HSTS
In your code, put in JSON.stringify to convert the object to a string from logging, console.log(JSON.stringify(event)); You can then see the logging in the terminal/command-prompt via sam local start-api Or have it log to a file with: sam local start-api --log-file logfile.txt
I am using the default HelloWorld example My code is: let response; exports.lambdaHandler = async (event, context) => { console.log(event); try { response = { 'statusCode': 200, 'body': JSON.stringify({ message: 'hello world' }) } } catch ...
How do I see the full logs of a local invocation of a Lambda via SAM?
I would bet good money that the problem is to do with the fact that when you run that script by hand you run it from the directory whereHitURL.jaris -- whereas when cron runs the process, it would probably use the root directory as the current directory. As such I'd change it to include the full path toHitURL.jarin the...
I have written a shell script which has a Demo class having main method.Now when I am running my shell using./file.sh.It works fine.But now I have configured a cronjob to execute that file every 5 mins usingcrontab.But this is giving me aclassnotfoundExceptionException in thread "main" java.lang.NoClassDefFoundError: c...
ClassnotFoundException while running a java class through crontab using unix?
The sonar.exclusions / sonar.global.exclusions are not what you want to use. They are for excludingfilesfrom the analysis notrules. To exclude a rule specifically for a project you must define, in SonarQube, a quality profile and remove that rule from the profile, and assign that quality profile to your project. You ca...
I have a Console project and want to disable the ruleConsole logging should not be usedon it. But it seems MSBuild don't get theses exclusions (anything found in Agent logs, or .sonarqube logs) added in project configuration (sonar.issue.ignore.multicriteria).Rule Key Pattern:csharpsquid:S2228File Path Pattern:**/MyPr...
SonarQube sonar.issue.ignore.multicriteria and MSBuild
0 You should use -d flag that means detached instead & sign for last docker command. So full command will be: docker run -d --env-file=Timmy-SchoolSimplified/.env spaceturtle0/ss-timmy Share Improve this answer Follow ...
so I'm currently trying to make GitHub Actions/CI SSH into my VPS and run a docker image. Although the main problem is that the job doesn't finish up after running the final command. This is my YML file: name: SSH & Deploy Image on: workflow_run: workflows: ["Timmy Docker Build"] branches: [ main ] types...
Running a Docker Image via SSH Github Actions
You can't have multiple profiles for the same project. But if you want this, you just have to create your own custom profile which has all the rules defined in your profiles. For Example, let's say you have quality profile "Android Lint" and "Sonar Way", and if you wan't to check your project against both these profile...
Is there a way to have several profiles for one project in Sonar ?For example we have a Java EE project with : EJB and WebApp (JSP and Back Bean). We have 3 quality profiles: Java, Web and XML. To "execute" these 3 profiles on our project we run three sonar analysis and create 3 result projects in Sonar.We use views to...
Several quality profiles for one project
I found an interesting resourceherethat helped me to achieve my goal above.so the it's possible to define additional user data section in the instance group configuration.In my case I was about to enable Nvidia runtime which I achieved as follows:spec: additionalUserData: - content: | #!/bin/sh sudo sed...
Closed.This question does not meetStack Overflow guidelines. It is not currently accepting answers.This question does not appear to be abouta specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic onanother Stack Exchange site, ...
Configure a new kubernetes node automatically [closed]
As noted in the comments, all you have to do is remove the rules from your profile or edit them to lower their priority. You need the Global Administer Quality Profiles permission to do that. Once you're logged in with that permission, go to the Rules interface, search for a rule you want to deactivate, select the rule...
We recently started usingSonarQube. We have found some rules that are suggested by SonarQube but we want to ignore them or give them a low priority and even configure the time suggested by SonarQube. For e.gWe want to avoid the rule (and/or configure the priority and time suggested by SonarQube) forDocument this public...
How can we ignore some SonarQube rules in Java?
collect_listis not better than just callingcollectin your case. Both are incredibly bad idea for large datasets. and have very little practical applications.Both require amount of memory proportional to the number of records, andcollect_listjust adds overhead of shuffle.In other words - if you don't have a choice, and ...
I have a large dataset of 50 million rows with about 40 columns of floats.For custom transformation reasons, I am trying to collect all float values per column usingcollect_list()function ofpyspark, using the following pseudocode:for column in columns: set_values(column, df.select(collect_list(column)).first()[0])Fo...
"java.lang.OutOfMemoryError: Requested array size exceeds VM limit" during pyspark collect_list() execution
Your date format must betar --newer 20120408for newer files than Apr 8, 2012. No "-"!Or you can useTWODAYSAGO=`(date --date '2 days ago' --rfc-3339=seconds)` tar -cz --newer-mtime="${TWODAYSAGO}" -f bakfile.tgz /dir_to_backup
I wanted to make an incremental backup with tar. I have made a full backup on 2012-04-08 and later I wanted to backup all files created or changed after that date.So I did something like this:cd /directory/I/wanted/to/back/up tar --newer 2012-04-08 -cvnf "/backup/dir/$(date +%F).tar"After a while I realised that tar is...
Is tar ignoring --after-date option?
Sequential addresses refer to sequential bytes in memory.An address that is "4-byte aligned" is a multiple of 4 bytes. In other words, the binary representation of the address ends in two zeros (00), since in binary, it's a multiple of the binary value of4(100b). The test for 4-byte aligned address is, therefore:if ( (...
My question has two parts.First, as a newbie to this address space, I would like to know what is the meaning of memory alignment of an address. I Googled about it but wanted to ask this question here as well since I found answers here very useful.The second part of my question is related to alignment and programming: h...
How Do I check a Memory address is 32 bit aligned in C
To get the maximum amount of memory that can be allocated, I would actually allocate memory in ascending order, using realloc to double the allocated memory each time until it fails. However, there's no guarantee that the allocated memory is actually in physical RAM, which may matter depending on what you plan on usin...
I have a MFC application written in VC6. I would like to temporarily allocate possible maximum memory for certain operation. In order to do that, I need to find how much memory I can allocate at current state of the program. Is there any useful function or code to do this? Of course, I can try to allocate memories mul...
how to find maximum amount of allocatable memory in current state of program
##[error][SQ] API GET '/api/server/version' failed, error was: {"errno":"ETIMEDOUT","code":"ETIMEDOUT","syscall":"connect","address":"","port":9000}To solve this issue, you need to add the correct service tag to the VM NSG rule if you are using Microsoft-hosted Agent.Refer to the following steps to confirm the service ...
I have a simple setup with a single Azure VM running sonarqube, and an NSG attached to the NIC with a public IP. I have been unable to get the NSG inbound rule to work with either Service Tag 'AzureDevOps' or 'AzureCloud'. Everything else in the NSG has been configured at this point to either Any or * other than the de...
Cannot get Azure DevOps connected to SonarQube on Azure VM
0 If your users are Enterprise Managed Users, you now (Sept. 2022) have the option of: Automatic Single Sign On for Enterprise Managed Users – Public Beta Enterprise administrators can now choose to redirect signed-out Enterprise Managed Users to their company's single sig...
As a company, is it possible to mandate SSO for everyone who works for me? i.e anyone who registers with an @companyname.com email has to go through my Google SSO regardless if they're logging into and pushing code to a repo not maintained by my organization?
GitHub SSO - Enterprise Mandate
0 On Linux, UNIX, and Windows platforms, the HADR feature has been a popular approach for mirroring a DB2 database to an identical standby database since it was first introduced many years ago in v8.2. Starting with DB2 v10.1, HADR can mirror the primary database to up to t...
I'm seeking for some guidance in performing mirroring with a witness in DB2. I've been searching for hours... I've found some information about split mirroring, but I've found nothing about how to configure a witness in DB2. Please help! Regards
database mirroring with a witness db2
You can simply pass nothing to the main class, to use the API without authenticating.from github import Github g = Github() r = g.get_repo("REPO_USER/REPO_NAME")ShareFollowansweredJun 3, 2021 at 17:32ZeromikaZeromika34722 silver badges1313 bronze badges1Thank you, this is what I was looking for–BloodLordJun 3, 2021 at...
I couldn't find anything online so that's why I'm posting it despite the question being simple.I'm writing a script that download an application's binary and puts it in a file. I would like the use the PyGitHub API to retrieve some information. The problem is that I don't want to enter any credentials when accessing Gi...
How can I download files from a Github repository without logging in?
We have already changed the default Regex for this rule to allow underscores. The next version of the C# plugin will use that. Until then you can change the Regex yourself to^[A-Z][a-zA-Z0-9_]*[a-zA-Z0-9]$.
The Sonar rule csharpsquid:S100 (Method name should comply with a naming convention) is thrown also for event handlers that are generated by Visual Studio, something like:protected void Page_Load(object sender, EventArgs e) { DoIt(); }Is it possible to ignore this rule for event handlers as they are auto-generated?
Method names for event handlers S100
FromApache's docs on mod-rewrite:Modifying the Query StringBy default, the query string is passed through unchanged. You can, however, create URLs in the substitution string containing a query string part. Simply use a question mark inside the substitution string to indicate that the following text should be re-i...
I have url for ex:http://www.demo.com/a/abcdAnd I use this htaccess to post this url's valuesRewriteRule ^a/(.*)$ /details.php?sef=$1 [L,NC]But I need to get new parameters posted the url For ex:http://www.demo.com/a/abcd?id=123&qu=11So how can I get theidandquvariables values via this url without do any changes in UR...
post parameters to htaccess sef link
A domain name (for use as websites etc.) should point at an IP address. There are several kinds of DNS records. Records for pointing at IP addresses are A records, as you have in 1). Other kinds are MX records for mail servers, for example.If you want to create a subdomain, you want to create another A record, so simpl...
I created a domain at freenom.comhttps://i.stack.imgur.com/an4tz.jpgwhich points to my app at Digital ocean droplet. and next day I created this domain I tested and it works. that domain is 1) at printscreen above.Now I want to create several subdomains, as I know that is possible to make:site.com - domain dev1.site.co...
How to create a subdomain at freenom.com?
I installed Cygwin (Linux Platform on Windows) to execute the script.
I need to setup a windows authentication in Kubernetes. And to configure GMSA in K8s for pods and containers in windows, I came across this link:-(https://kubernetes.io/docs/tasks/configure-pod-container/configure-gmsa/).This documentation has a step which confirms to “Install Webhooks to validate GMSA users”. To follo...
Kubernetes - Configure GMSA for Windows Pods and containers
There is no such thing as the smallest unit of Kubernetes is a pod - which is either started directly (create pod) or controlled by another resource, eg. replicaset, job, cronjob.The pod creation time should be fairly small if you have the image already present on the worker node. Are your workers terminating after eac...
As of now, I'm spinning up individual K8s jobs for a specific processing task. Some of these tasks require significant CPU/memory but other are fairly simple and could easily be accomplished with in-memory processing. A simple task that takes a few milliseconds in-memory is much slower in comparison when running as a K...
Is it possible to have a worker pool for Kubernetes Jobs to avoid pod creation time?
2 I had same issue with boto3 s3 client in my client class and moto in my pytest. I resolved it by wrapping boto3 client into a singleton: This is my client code hello_world/app.py class Singleton(type): _instances = {} def __call__(cls, *args, **kwargs): if cls not i...
For lambda, it's best practice to initialise dependencies outside the handler. I am creating a simple python function that works like the blueprints: import boto3 s3 = boto3.client('ssm') def lambda_handler(event, context): # some code here And the test from lambda_function import handler # Option 1 import l...
Testing Python AWS Lambda boto3 initialization
Yes. as far as the certificate is concerned there is no relationship between different servers. You could have five different servers with five different certificates and as long as each one of them is valid there wouldn't be a problem. (itmightcause a problem if the certificate changes mid-session)You should however b...
My company has a SSL certificate for *.mycompany.com. We use this certificate for many critical services.Marketing team owns www.mycompany.com website (WordPress) and wants a SSL certificate for it.Can we issue a specific certificate for www.mycompany.com?
Can I have different SSL certificates for different subdomains?
You can use this not (yet) perfect script: if [ "$#" -ne 3 ]; then echo "Illegal number of parameters, usage: script file lineStart lineEnd"; echo "Example : test.sh .travis.yml 12 16"; exit 255; fi branch=`git rev-parse --abbrev-ref HEAD`; repo=`git config --get remote.origin.url | sed 's/\.git//g'`; ec...
I'm often sharing link to my code with my colleague using Github and Gitlab cool line anchor feature (ie: https://github.com/Phonations/Joker/blob/master/.travis.yml#L12-L16). I'm looking for a way to quickly jump from my file to it's web version. Does anyone know a trick or an extension to perform this? If it is poss...
Switch from my Xcode versioned file to its github/gitlab web version
got it. the problem was permission error to /dev/video0 just changed /dev/video0 to rtsp address of camera
I am having an issue with nginx-rtmp-module exec ffmpeg command. i have followed the example on www.github.com/arut/nginx-rtmp-module /home/junaid/bin/ffmpeg -f video4linux2 -i /dev/video1 -c:v libx264 -an -f flv rtmp://127.0.0.1:1935/myapp/mystream the command is working fine in terminal Config file: rtmp { server ...
nginx rtmp-module can't execute ffmpeg
In case anyone google to here:After spent some time to research, I found it is quit easy.First, we have to forbidden all server with wild match:server { listen 80; server_name *.example.com; return 404; }Then you can define all known subdomains as servers.ShareFollowansweredDec 13, 2023 at 15:04mtdcy.chenmtdcy.ch...
I have one valid subdomain on my site; it and the main domain work. But I was wondering, what if someone enters a bogus subdomain?I have a rule like this:# throw away all other subdomains server { server_name ~^(.*)\.foo\.to$ ; return 301 https://foo.to/notfound; }I thought this would fix the problem but instead t...
How to redirect non-existing subdomains in nginx to 404
4 It's quite easy to overload operator new() in the class. The global one can be then called using :: to specify global namespace as in ::operator new(). Something like this: class ClassX { public: void* operator new( size_t size ) { // whatever logging you ...
I have a class classX and would like to know how much how much memory all of the instances of this class use. Every new instance is created using new classX Is there a way to do this without modifying source code (ie using tools like valgrind)? And what methods can I use to do this by modifying the source code (I can'...
Finding memory usage of a single class in c++
This feature seems tonot be available for GitHub wikis.I have also tried the empty_Sidebar.mdfile as suggested by lucifurious, but all it does is create an empty (blank) sidebar below the "Pages" one.
It is easy to define a sidebar on GitHub's wiki page by creating_Sidebar.mdfile in the root and the sidebar will be automatically appended to every page. But what if we would like to hide it on certain pages, is there any way to achieve this?
How to hide a sidebar on GitHubs wiki page for certain pages?
There are basically two meaningful metrics: latency and throughput.Latency answers the question "how long do I have to wait after agit pushuntil the build (or whatever else) is finished"? and throughput answers the questions "how many builds per hour or day can I make?".A system with a high latency can still have a hig...
I need to measure the performance of my Continuous integration environment in GoCD, docker and GitHub repository. Could you please recommend me ant metrics for the evaluation. It is also be great if you recommend me how to apply those metrics. Thank you in advance
Performance measure of Continuous Integration in GoCD
Ok its not 100% bulletproof since HTTP REFERRER can be spoofed.Try this rule:RewriteCond %{HTTP_REFERER} !^http://(www\.)?yourdomain\.com/ [NC] RewriteRule ^process/login\.php$ - [F,NC]
Not sure if this is possible, but...Lets say i have process/login.phpif a user visits that page directly i have a check to see if the user has come from posting a form etc. If not, they get 404'd.However i am now building up plenty of process scripts etc and they all require different checks as some are just included e...
Stop direct access to all files in a folder, but allow ajax requests
Try scaling back to a minimal configuration without SSL and ensure things work for 2 domains first:server { listen 80; server_name example.com; return 200 "example.com"; } server { listen 80; server_name 1.example.com; return 200 "1.example.com"; }$ curl http://example.com/ example.com $ curl http://1.exa...
I have the following configuration in my sites-available (with a symbolic link to enable):#subdomain site server { listen 443 ssl http2; server_name dokuwiki.[censored].org; root /var/www/html/dokuwiki; index index.php index.html; include /etc/nginx/templates/ssl.tmpl; location / { try...
nginx subdomain with different site than root
finally i find the solution herehttps://newbedev.com/invalid-ssl-certificate-when-pushing-to-git-serverAfter reading and making the settings, you should dogit config --system http.sslCAInfo "C:/Program Files/Git/mingw64/ssl/certs/ca-bundle.crt"and in the root of your project that you are uploading to gitLab you should ...
I'm trying to push my first project in gitLab but i get this error "fatal: unable to access 'https://.git. : SSL certificate problem: unable to get local issuer certificate"
GITLAB SSL certificate problem: unable to get local issuer certificate
2 You need some basic notions about git. When you are at your second step: (use "git add <file>..." to update what will be committed) (use "git checkout -- <file>..." to discard changes in working directory) modified: public/javascripts/app.js git is telling you that t...
I am attempting to unstage two files. When I type git status my message is: On branch master Your branch is up-to-date with 'origin/master'. Changes to be committed: (use "git reset HEAD <file>..." to unstage) modified: public/javascripts/app.js new file: views/templates/caffeineTable.html I use...
unable to move files from unstage with git
You could use restart policies. I do not know how exactly ansible creates containers. But usually you could add a parameter on command line like: docker run --restart=always ...
Currently we use the combination of Ansible + Ansible Tower to deploy and run our applications in Docker containers. However, I was wondering how to automatically run the deployed containers after a system reboot. What would be the way to go here? The only solution we came up with is to automatically request the Tower...
How to start Docker containers on boot, managed with Ansible (Tower)
1 It seems that you need to add only one extra line to your Dockerfile: RUN systemctl disable php-fpm that will disable php-fpm from being spawned from your systemd process, your process will be definitely started by supervisor and that will make the image to be compatible...
I have simple Docker image built over ubuntu with dummy laravel php application. I use supervisord to start nginx and php-fpm. So far so good, everything works fine in my local machine or in any place where docker executable exists. I'm trying to run same image in the Jelastic environment but I'm getting supervisord e...
How to properly deploy Docker containers with supervisord to Jelastic
Do I need to enable Prometheus plugin on the three servers ?Since you are deploying Kong using Hybrid mode deployment, applying the Prometheus plugin on Control plane (via Admin API or Kong Manager). After applying the plugin onto Control plane the Data plane will be configured to expose Prometheus endpoint.The metrics...
I am using Prometheus with Grafana to monitor KongAPI Gateway by enabling Prometheus plugin on Kong.Our Kong architecture uses a dedicated one server for Control Plane and another two servers for Data Plane.Do I need to enable Prometheus plugin on the three servers ?Are Metrics of the servers different so that I need t...
Monitor KongApi using Prometheus
If the sonarcube error is your title ("Associate a valid label to this input field sonarqube"). This may be warning you of an accessibility issue as there is no label directly associated with your input for the sake of screen readers.Try either adding a for attribute to your label tag to target the input<label for="lan...
Sonarqube found a bug in this line of code:<div class="dropdown-language"> <label>{{'GENERALE.LINGUA' | translate }}</label> <select #langSelect (change)="translate.use(langSelect.value)" class="lang-style" id="language"> ---> **HERE** <option *ngFor="let lang of translate.getLangs()" [v...
Associate a valid label to this input field sonarqube
0 On your victor-order-error-code branch, run git revert SHA where SHA is the SHA hash(es) of the commits that you would like to remove. You can specify a range of SHA hashes and it should be your entire victor-boos-833 branch since master. See http://sethrobertson.githu...
I created branch victor-boos-833, worked on it, then push to origin as a new branch, then made a pull request. I then created another branch victor-order-error-code from master (I think), worked on it, then push to origin as a new branch. The problem is, changes in victor-boos-833 which has not been merged, were also ...
Git pull request multiple branches but other branch'es commits were added
If you are using index documents, where this... http://bucket.example.com/foo ...actually displays the document at foo/index.html then your workaround will be to avoid the redirect that S3 does prior to displaying the index page. For a link pointing to the page at /foo, you should actually create the link pointing to...
I am hosting a static website on Amazon S3. Some of my client-side javascript parses the query strings to control the HTML. This works fine locally, but on the S3-hosted version, the query strings seem to get dropped from the request. My motivation for using query strings is that I want to be able to pass state betwee...
How can I get query strings in my Amazon S3 static website?
AddQSAflag in last rule:RewriteRule ^$ /splash [L,R=301] RewriteCond %{HTTP_HOST} ^example\.com$ [NC] RewriteRule ^ http://www.%{HTTP_HOST}%{REQUEST_URI} [R=301,L,NE] RewriteRule ^([a-z0-9-]+)/?$ /index.php?cat=$1 [L,NC,QSA]QSA(Query String Append) flag preserves existing query parameters while adding a new one.
I managed to put together .htaccess like this and it is working fineRewriteRule ^$ /splash [L,R=301] RewriteCond %{HTTP_HOST} ^example.com RewriteRule (.*) http://www.examle.com/$1 [R=301,QSA,L] RewriteRule ^([A-Za-z0-9-]+)?$ /index.php?cat=$1 [L]This redirects things likeexample.com/onetohttp://www.example.co...
.htaccess multiple variable rewriting
Helm uses KubernetesDeploymentwith a different terminology than Docker. You'll want to define:commandin Helm forentrypointin Docker Compose (seethis post)workingDirin Helm forworking_dirin Docker Compose (seethis post)For your example it would be:... containers: - name: checklist ... command: ["dotnet", "Checkli...
I have the followingdocker-composefile and I don't get how I can set theworking_dirandentrypointin the helmdeployment.yaml. Does someone have an example on how to do this?docker-composeversion: "3.5" services: checklist: image: ... working_dir: /checklist entrypoint: ["dotnet"...
How to configure docker entrypoint in Helm charts
You can export tables and queries with fbexporthttp://fbexport.sourceforge.net
I would like to backup a Firebird database but exclude certain tables from the backup, is this possible?If not, I'd like to make a copy of the Firebird database while it's running (without doing a backup followed by a restore)Neither gbak nor nbackup seem to support things like this, and yet we have a piece of software...
Partial/Selective backup/copy of a Firebird database
EXPOSEinforms Docker that the container listens on the specified network ports at runtime but does not actually make ports accessible. only-pas you already mentioned will do that:docker run -p :$HOSTPORT:$CONTAINERPORTOr you can opt for a docker-compose file, extra file but also do the thing for you:version: "2" servic...
I am writing a simple docker file for a golang and I'm still getting familiar with docker so I know what I want to do just don't know how to do it:What I have right now (below) is exposing port 8080, but I want to expose port 80 but forward that to port 8080.I know that I can do it viadocker run -pbut I'm wondering if ...
How to bind docker container ports to the host using helm charts
As of 2015-08-07, Elastic Beanstalk does not support Aurora. I received this message from AWS Support:Hello, At this time it is not currently a supported option for use with Beanstalk, though we are working to expand where and how it can be integrated. Please stay tuned to our release blog for any updates in the nea...
In the Oregon region, Elastic Beanstalk let's me select "aurora" from a list of RDS databases. However, when I try to launch the application, I get this flash message:DBEngine: Invalid option value: 'aurora' (Namespace: 'aws:rds:dbinstance', OptionName: 'DBEngine'): Value is not one of the allowed values: [mysql, oracl...
Does Elastic Beanstalk support Amazon's Aurora DB?
In git (and other distribution version control systems or DVCS), there is the concept of a 'local' and a 'remote' repository. You have only commited to your local repository, and not yet 'synced' to the remote repository (GitHub) - hence your changes are not yet displayed on GitHub.Specifically in Git you create acommi...
I am committing changes from android studio into github and message appears that " succesfully committed 2 files ... ". But i cant see even a single change appearing on my repo on browser when I refresh it. Is there a way to know where those changes were committed?
Changes commited from Android Studio
You're calling free() on this pointer: b[i] = "baz" That pointer wasn't allocated with malloc() (or calloc(), realloc(), etc).
Consider the following code: #include <stdio.h> char** baz_alloc(int size) { char ** b = malloc((size+1) * sizeof(char*)); for (int i = 0; i < size; i++) b[i] = "baz"; b[size] = NULL; return b; } void baz_print(char** baz) { char** b = baz; while (*b != NULL) { printf("%s...
Failing freeing a 2d dimensional array dynamically allocated
The heart of the issue is essentially a circular dependency. Constructing the RTree causes the rtree<...> template instantiation which includes a typedef node_pointer = allocators_type::node_pointer, which triggers the instantiation of allocators_type, i.e. detail::rtree::allocators<...>, which has a base class of det...
I am trying to use Howard Hinnant's stack_alloc with boost rtrees, as in the following example: #include "stack_alloc.h" #include <boost/geometry/index/rtree.hpp> using NodePoint = boost::geometry::model::point<double, 2, boost::geometry::cs::cartesian>; using Linear = boost::geometry::index::linear<8, 2>; using RTre...
Hinnant's stack allocator with boost rtrees: compilation failure
To be able to access Redis Server from Grafana Cloud it should be exposed to the Internet as Jan mentioned.If you run Grafana in Docker container it should be started in the host network mode (https://docs.docker.com/network/host/) to be able to access it from other devices.If something is lacking or not clear in the R...
I'm trying to access my redis database via Grafana Cloud on my laptop. The database is a redis container working as a cache on a different device (pi). Accessing the Redis database via Python script on my remote device is no problem but trying to connect to it via Grafana (using Redis Datasource Plugin) doesn't work as...
Accessing remote redis server with Grafana
0 May be you should try to run it with --privileged flag. docker run -ti --privileged yourimage But make sure that you know what you are doing. You should also read docker-tips-privilaged-flag Share Improve this answer Follow ...
I have run free -h and see that I have 29G of swap space. total used free shared buff/cache available Mem: 15G 6.9G 8.8G 17M 223M 8.9G Swap: 29G 2.0M 29G I have also enabled 100 swappiness. $ sudo sysctl vm.s...
Make docker build --memory-swap=20g use the available swap space?
First, do note that in the link you providethat rule is deprecated:This rule is deprecated, and will eventually be removed.Second, I'm mpt sure there's a way to shut up sonarqube (which is probably why the rule's getting removed, but I don't really know the tool), it's really more of a warning sign telling you to be ca...
x = input("Enter name") if x.lower() == "keshav": print(x)When I run sonarqube on this code then I am getting security Hotspots issue.You can check more related to this issuehttps://rules.sonarsource.com/python/RSPEC-4829anyone have any idea how to fix this security issue of sonarqubeI have tried using cgi.escape(...
How can I sanitize the standard input in python to fixes the sonarqube issue
It might be too late to you on the update for this query.I had found the below approach to get the exact value. It was not quiet easy one with any direct environment variable or with action. Basically use the actions GitHub API with run_id.Below is a snippet of GitHub-action step.- run: | URI="https://api.github.c...
I am using build.cake file tasks with PowerShell script on the windows machine for GitHub CI. Am trying to get the check_run_id via GitHub Actions in the workflows .yml file. Is that possible?i.e.https://github.com/Siddharth/my-project/pull/15/checks?check_run_id=2508655272I need the id(2508655272) value in workflows....
How to get the GitHub CI job build id (check_run_id)?
I was able to solve it.What i was missing is in the query part in graph i missedgroup by tag name.In my case its group bytag(label). and that solves the issue.
Im trying to getthroughput/secofeach samples(i have in myjmeter test) ingrafana dashboardusinginfluxdb.I'm usingtemplatingto configure .screenshot showing template settingsMy query is--SELECT count("responsecode") FROM "samples" WHERE "label" =~ /$label/ AND "status" =~ /$status$/ AND $timeFilter GROUP BY time($int...
How to get more than one graph of samplers (which i have in my jmeter) in a graph of grafana using templating
Try first to remove the old certificate from your list in /etc/ca-certificates.conf:sed -i 's#mozilla\/DST_Root_CA_X3.crt#!mozilla\/DST_Root_CA_X3.crt#g' /etc/ca-certificates.conf update-ca-certificates --fresh
I'm running debian 9 in my docker container and today I was not able to update root certificate for Lets encrypt inside container. On host Ubuntu 20 everything worked without any my intervention, but on my debian container I was not able to get new root Lets Encrypt certificate, only recreation container from scratch h...
Refresh Lets encrypt root CA in docker container
Check: Supported Resources for Tag Editor Tagging (Console) for services and resources that can be tagged using AWS Console.
Is there any easy way to find the consolidated AWS Services which has tags? I am currently looking each service in AWS documentation and finding out, which is tedious. For example - services like ec2,s3,lambda,dynamodb - has tagging feature, so that resources can be identified by tags(assuming they are tagged). Servic...
is there any source for finding out aws services which support tags?
+50Certbot uses a configuration file to keep track of all the settings it uses, or at very least a record of the non-default ones. Seeherefor information on that. However, updating certbot itself might be enough as I believe recent versions have been disabling 1.0 and 1.1 by default.Honestly, I would also remove TLS ...
My current cerbot version running is:certbot --version certbot 0.27.1My ngnix has TLS v1.0 but I do not want that running anymore:cat /etc/letsencrypt/options-ssl-nginx.conf # This file contains important security parameters. If you modify this file # manually, Certbot will be unable to automatically provide future sec...
Certbot version update in NGNIX to remove TLS v1.0 on MacOs
Using aws and jq from the command line (tested with bash on mac): export CS_DOMAIN=https://yoursearchdomain.yourregion.cloudsearch.amazonaws.com # Get ids of all existing documents, reformat as # [{ type: "delete", id: "ID" }, ...] using jq aws cloudsearchdomain search \ --endpoint-url=$CS_DOMAIN \ --size=10000 \...
I have an AWS CloudSearch instance that I am still developing. At times, such as when I make some modification to the format of a field, I find myself wanting to wipe out all of the data and regenerating it. Is there any way to clear out all of the data using the console, or do I have to go about it by programatic mea...
How to clear all data from AWS CloudSearch?
A pull request is a way to 'commit' to a repository in which you don't have writing permissions. The maintainers of that repository will check your request and decide if they either want to merge it with your code or leave the original as it is.
As I have read that commit means we have updated some code which has bugs and commited that code from where the issue is raised. But what is a pull request.
What is the difference between commits and pull requests
Debugging should be similar just like we have it in Dot net core. In dot net , we used to debug something like thisSetting and using breakpoints for debuggingIf Visual Studio 2017 is still connected to your dev space, click the stop button. OpenControllers/HomeController.csand click somewhere on line 20 to put your cur...
How to debug the code written in python in container using azure dev spaces for kubernetes ?
Azure dev spaces to debug python code running in kubernetes pod
It looks like they are trying to optimize things - by using docker caching and layering.Splitting off restore from build/publish is a good idea. Your code changes often, but your project files and dependencies change little. Docker uses caching for things that don't change. If you can separate your project file from th...
VisualStudio generates Dockerfile for .NET Core like this:FROM mcr.microsoft.com/dotnet/aspnet:5.0 AS base WORKDIR /app EXPOSE 80 EXPOSE 443 FROM mcr.microsoft.com/dotnet/sdk:5.0 AS build WORKDIR /src COPY ["src/Sandbox/Sandbox.csproj", "src/Sandbox/"] RUN dotnet restore "src/Sandbox/Sandbox.csproj" COPY . . WORKDIR "...
Why restore and build are needed before publish in dockerfile autogenerated for .NET Core by VisualStudio
You may just use video-sink property of playbin:gst-launch-1.0 playbin uri=https://www.freedesktop.org/software/gstreamer-sdk/data/media/sintel_trailer-480p.webm video-sink=xvimagesink gst-launch-1.0 playbin uri=https://www.freedesktop.org/software/gstreamer-sdk/data/media/sintel_trailer-480p.webm video-sink=nv3dsinkS...
Assuming that there are 2 kinds of streams, one is only video stream (without audio), the other is video with audio. We know that playbin with a uri can play them all even if we dont know what kind of stream we get, but is there any pipelines that use xvimagesink or nv3dsink (not autovideosink etc.) that can receive bo...
A general gstreamer pipeline use xvimagesink to pull rtmp stream with/without audio
Not only you have to mount it in your VirtualBox, butyou also have to instruct, in your boot2docker TinyCore session that you want that folder visible(once you have done adocker-machine ssh yourMachine):mount -t vboxsf -o uid=1000,gid=50 your-other-share-name /some/mount/locationI know that you are usingdocker-machine,...
Using docker client, is there a way to share a folder in windows with a docker container without having to first share the folder via the Virtual Box VM.Have understood the need of having a double slash fromthisandthisRan the following command from the docker client for windowsdocker run -it -v //F/devfolder:/developme...
share windows folder (other than c/Users/) with docker container (using docker windows client)
1 Looks like its not supported :( see... How to SSH in to different containers in Multi Container Azure App Service and... Support SSH to specific container in multi-container setup Share Improve this answer Follow ...
I'm running an Azure Web app (containers) with custom container images. I've followed the steps to enable ssh into a container image and it works great when I'm only running a single container. But when I run the app as a multi-container app (with docker-compose file) with more than one container image I get the err...
Does Azure Web App for containers support ssh access when running a multi-container app?
If it's taking the input from stdin, make another script that will call your script with a pipe or a redirect.#!/bin/sh /foo/bar/my_command < my_inputYou can also launch it as a shell command in your crontab:0 * * * * /bin/sh -c "/foo/bar/my_command < my_input"
I have a run a script in cronjob. But that script is taking some user input.How can I handle such case?
How to give user input in cronjob script
Well this is a stupid answer but, the repository you are using doesn't seem to exist, did you created it ? If so and it's not public, is it really empty ? A git push --force might be of some use.
This question already has answers here: Issue pushing new code in Github (14 answers) Closed 4 years ago. I'm having some issues using Git. I already have a live site hosted on a D...
How to create Github repository from a site that's already live [duplicate]
After a lot of painstaking experimentation the problem has at last been solved.It seems that with the new iis8.5 server and Helion Ape I needed to put the rewrite rules into a web.config file instead of a .htaccess file.The new rule for the example I gave in the question is:<rule name="Imported Rule 6"> ...
I have been demented with this problem for 3 days and no matter how differently I search Google or read the Helion website I cannot understand what I am doing wrong. So, I do hope that someone will be kind enough to help me.I use a.htaccessfile to create search engine friendly URLS and it has worked well for several ye...
Is there something wrong with my rewrite code in .htaccess file
You are never releasing the animData. You need to add dealloc to your class: - (void)dealloc { [animData release]; [super dealloc]; } Similarly, you need to add a similar dealloc to AnimationData. On a related note, you need to retain and later release the string created in -[AnimationData init], what you are do...
I'm having problems with a leak in the init method of a class I have created. To keep it simple, I have the following (simplified) problem: ViewController initialises an instance of ClipData class which initialises an instance of AnimationData class which initialise a string ViewController: myClipData = [[ClipData ...
Memory leak in objective-c class init method
The problem is that step 13 of the tutorial reads:kops get secrets kube --type secret -oplaintextIt should instead be:kops get secrets admin -oplaintext
I followed this tutorial on Medium to deploy a Kubernetes cluster on AWS:https://medium.com/containermind/how-to-create-a-kubernetes-cluster-on-aws-in-few-minutes-89dda10354f4However, when I launch the Kubernetes dashboard I see the following errors:configmaps is forbidden: User "kube" cannot list configmaps in the nam...
Kubernetes: configmaps is forbidden: User "kube" cannot list configmaps in the namespace "default"
As far as I know, this is not possible.You can't sign-out a user that does not allow it.What you could do is torevokethe token, or redirect to sign out.Example of a request that needs to be sent to the revocation REST endpoint:curl -X POST --basic -u "<client id>:<client secret>" -H "Content-Type: application/x-www-for...
I've set upnginx ingress with oauthfor Kubernetes based off ofbitly'soauth2_proxy. There is metadata for anauth-urlandauth-signinpage, but I'm not sure if there is a way to configure logout (volentarily or by force). Obviously I'd like the ability to kick someone off if I need to.annotations: nginx.ingress.kubernet...
Nginx Ingress OAuth Logout (Kubernetes)?
Yes you can run multiple containers on a single host; docker is designed for exactly that.You cannot map two containers of different images to the same port number; you get an error response if you try. However, if your containers run the same image (e.g.2 instances of a webapp) you could run them as a service, and hav...
I have not done any practical with the docker and container, But as per my knowledge.As per the documents available online I did not get the details about the running two or more containers at the same time.Docker allows container to map port address of container to the host machine. Now, the question is can we run mul...
Can we have two or more container running on docker at the same time
Fixed! I replaced both lines with:RewriteRule ^shop$ magento/index.php [L] RewriteRule ^shop(.*)$ magento$1 [L]
I'm having great difficulty in rewriting my shop domain to my Magento subdirectory. Here's what I'm after:http://mydomain.com/shop/-> /var/www/magento/http://mydomain.com/shop/anyFile.php-> /var/www/magento/anyFile.phpThis is the ineffective rule I currently have in my /var/www/.htaccess file:RewriteCond %{REQUEST_URI}...
Rewriting one subdirectory to another
This seems to work (but I have to set the status code in PHP)RewriteEngine on RewriteCond %{REQUEST_URI} !^/static/.*$ RewriteCond %{REQUEST_URI} !^/media/.*$ RewriteRule .* down.php [L]and in down.php<?php header('HTTP/1.1 503 Service Temporarily Unavailable',true,503); ?>Any problems with this? My main concerns are ...
I was trying to install this .htaccess to notify my users of site maintenance. It seems the first [L] isn't working and the second rewrite is doing everything.How do you guys do site maintenance messages?RewriteEngine on RewriteRule ^s/down$ index.html [L] RewriteRule ^(.*)$ http://metaward.com/s/down [R=302,L]
Rewrite rule for "site down" pages
There is anownerfield that needs to be set. It is marked as optional, but necessary for this auth method to work, you need to set it to the name of your org. The end result provider config that worked for me:provider "github" { owner = "your-org-name" app_auth { id = var.app_id installation_id ...
I'm trying to setup team sync between Github and my id provider (Okta) using Terraform (v1.1.4). I've installed the Github provider as detailed inhere. I did create a Github App and installed on my org, the permissions I've tried:Read access to metadataRead and write access to administration, members, and organization ...
Terraform Github provider gets a 403 error on /user (using Github app auth)
Git determines if some file is binary by checking a few things in the file's content. These kind of files shouldn't probably be in your repo as they are auto-generated every time you compile your code. You can safely delete them and add and entry in your .gitignore file with specifying .class files so that you cannot ...
I've cloned a repo with git clone. git status says that it's all up-to-date. After that I open the folder with Eclipse and git status shows something different: $ git status On branch master Your branch is up-to-date with 'origin/master'. Changes not staged for commit: (use "git add/rm <file>..." to update what will...
Modified Java class (treated as binary files) immediately after git clone
4 There can be memory limits configured in either the web.config or machine.config, or both. In web.config the section is: <httpRuntime executionTimeout="3600" maxRequestLength="102400" /> In machine.config the section can also be the httpRunTime section, similar to: <ht...
I have an ASP.Net MVC application that allows users to upload images. When I try to upload a really large file (400MB) I get an error. I assumed that my image processing code (home brew) was very inefficient, so I decided I would try using a third party library to handle the image processing parts. Because I'm using T...
Is there any memory restrictions on an ASP.Net application?
SharedPreferences is a Class with access to the preferences that the PreferenceActivity generats. In other word the default preferences use the sharedPreferences class the same way you would if you did it manualy. The key it usus is of course the key you gave it in the preferences.xml Therefore you can use SharedPref...
I'd like to use the backup agent for saving user preferences: http://developer.android.com/guide/topics/data/backup.html In the article, it's explained how to backup SharedPreferences: http://developer.android.com/guide/topics/data/backup.html#SharedPreferences However it's not explained how to backup when a Preferenc...
Android 2.2 Data Backup: How to backup when PreferenceActivity is used?
In AWS your SSD disk is known as anEBS Volume.To update the volume size you would from the EC2 console want to find your current volume and right click on it. Then click modify volume and select your new size.If you haven't created your instance yet you can specify the size during the wizard.
I am new to AWS EC2 and I have just purchased a t3 medium reserved instance. I would like to add 100 GB of SSD storage to my instance and use it as the instance's primary hard disk. How can I do this? I did not see any option of adding and configuring the SSD disk when I purchased the instance.I have purchased a linux/...
Add SSD storage to AWS EC2 reserved instance
Click on advanced search on the search tab, that will do what you want:https://github.com/search?l=&q=CGLI&ref=advsearch&type=Code&utf8=%E2%9C%93But:You can't use the following wildcard characters as part of your search query: . , : ; / \ ` ' " = * ! ? # $ & + ^ | ~ < > ( ) { } [ ]. The search will simply ignore th...
When I am searching for example 'CGLI' it does not gives me back the matches for 'CGLIB':https://github.com/SquareSquash/java/search?utf8=%E2%9C%93&q=CGLIHow can I force to give it back?
How I could get back partial search results inside a GitHub repo?
You can use docker networkaliasesoptionsin your stack.ymlfile :version: '3.4' services: service-database: image: postgres:9.5.16-alpine networks: default: aliases: - my-special-hostif you want to use same host name from your host machine, add this to your/etc/hostsfile :127.0.0.1 my-s...
Problem setup:Docker Swarm3 services are running:service-somethingandservice-anotherandservice-databaseservice-somethingandservice-anotheruseservice-databaseto store data, likemongodb://service-database/blahblahWhat I want:Instead ofmongodb://service-database/blahblahI want to usemongodb://my-special-host/blahblahIdeal...
Docker - alias DNS entry (CNAME) for services
so I ended up taking a copy of the live server database and installing it locally, then I took a look at the console which was logging: Too many open files - Maximum file multiparts in content reached googling around for a bit knowing that the multi nested form of mine is the cause of the problem, I decided to incre...
We created a multi-nested form in Rails 5, the forms having image objects, many fields. after having 200+ items in a single form and trying to save the form, we started getting those errors. An unhandled lowlevel error occurred. The application logs may have details. there is nothing logged in the error logs, nginx ...
Huge multi nested form failing with "An unhandled lowlevel error occurred. The application logs may have details."
If the target address and the server at that address is setup to use https the javascript get will be encrypted automaticly. Theres is nothing done manually in javascript whatsoever
Closed. This question needsdetails or clarity. It is not currently accepting answers.Want to improve this question?Add details and clarify the problem byediting this post.Closed9 years ago.Improve this questionIs it possible to send a SSL certificate got from a textbox/textarea to a https GET query?I need to make a GET...
Send SSL certificate in javascript GET query [closed]
Looks like permission issue.Does the Windows VM has write permission for the Linux Container?Caused by: java.nio.file.AccessDeniedException: /home/jenkins/? \ Caused by: java.lang.IllegalArgumentException: Root directory not writable: ?/.jenkins/cache/jarsIn Linux Container did you verified above path exists or not?
Using Master on Windows VM whereas tried to spin container in linux container. Kindly help to resolve this.Even some idea/ guidance would be of great help.hudson.remoting.ChannelBuilder withJarCacheOrDefault WARNING:Could not create jar cache. Running without cache.java.io.IOException:Failed to initialize the default...
Jenkins Master - Docker Slave: Build shows error
What is the storage class that you use? Storage class hasvolumeBindingModeattributes that controls how PV is dynamically created.ThevolumeBindingModecould beImmediateandWaitForFirstConsumermode.For checking the storage class you can dokubectl get storageclassorkubectl describe storageclass. The default storage class wi...
I have a Kubernetes cron job in AWS EKS that requires a persistent volume, so this is roughly what I have:apiVersion: v1 kind: PersistentVolumeClaim metadata: name: pvc-{{$.Release.Name}}-tmp spec: accessModes: - ReadWriteOnce volumeMode: Filesystem resources: requests: storage: 10GiThen it's mou...
Helm timeout for cronjob with a persistent volume
try this (*/30 08 * * Mon) and for any more queries visithttps://crontab.guru/#*/30_08_*_*_Mon
I am trying to run a cron job using node-cron every Monday at 8:30 so I use "30 8 * * Mon" which never runs (I also used "30 08 * * Mon" to be sure). After a bit of troubleshooting, I have seen that "30 * * * Mon" does work and runs on the 30th min of every hour. Can anyone help me figure this out, please?
Why is my hour parameter not working in node-cron
Yes, it does. However, the URL bar in github web page no loner have git protocol for copying the URL to clipboard. You have to replace https:// with git:// by yourself.Also, git protocol use port 9418. Make sure you can use this port.git protocol is not encrypted, but it is faster than https.git protocol is read-only i...
Is it possible to use such configuration? Or there are no advatages between ssh:// and git://?
Does GitHub support git:// protocol for pull?
Fellas, I believe that I have managed to do the task:#!/bin/bash service="ssh" if (( ! $(sudo service ssh status | cut -d" " -f 3 | cut -d"." -f 1) == "running" )) then sudo /etc/init.d/ssh restart fiI have changed the LogLevel to Verbose, I hope the next time I will track more clues regarding the failure of t...
I am trying to create a Bash .sh script for a cronjob that starts the OpenSSH server if it is down or failed.Last night the SSH server was down and when I tried to access it today (from work) the connection was refused ofc. No traces in the /var/log/messages for the failure.So the question is -how to determine is sshd ...
Starting sshd automatically if it is down/failed
Your configuration should be in a stream block You don't need server_name localhost:389; You are including the configuration from /etc/nginx/conf.d folder which is included inside http block in main nginx.conf file. The stream block should be at the same level as http block. Check the /etc/nginx/nginx.conf for the in...
We have one "OpenLDAP" server with port 389 currently active,using nginx we want to proxypass this TCP port 389 to TCP based ingress. can any one please share the nginx.conf detail for this. So far, left with incomplete as per below, upstream rtmp_servers { server acme.example.com:389; } server { listen 389; ...
nginx proxypass TCP 389
I actually found that I was able to connect, and previously I was failing because not following some steps quite hide in the AWS Codecommit docs:I am not sure if giving the following at prompt are necessary steps, but I guess they do not hurt:At terminal provide the following:aws configureyou will be prompt to provide:...
I have set 2 repositories in AWS Codecommit and using before SourceTree from Windows, setting up access via HTTP protocol was and everything was working fine.Right not I am migrating the development of a project on a machine where I do have installed ubuntu and where I choose to use gitkraken.I was able to git clone in...
clone AWS codecommit repo via HTTP
As you know already, you are creating an endless amount of threads without properly stopping the previous one. To wait for a thread to terminate there is a .join() method. Here is the documentation for the Thread module:docs.import threading def worker(): if nextJobActive() and number_of_active_threads<5: ...
I have memory leak, but I can't find a way to solve it. I think the reason is for that because I use threads and don't stop/kill it in a right way.I have following method:import threading def worker(): if nextJobActive() and number_of_active_threads<5: t = threading.Thread(target=startThread, args=(my_lis...
How to handle out-of-memory error while using threads in python
Thanks to @valyala for pointing me in the right direction to solve this problem.Firstly, when defining therisk_queryquery variable in Grafana, use.*as the "Custom all value".Second, use{label_risk=~"$risk_query"}instead of{label_risk="$risk_query"}:sort_desc( sum( max(kube_namespace_labels{label_risk=~"$risk_query"}) b...
I have a PromQL query in Grafana that returns the CPU usage of all Namespaces in my Kubernetes cluster, aggregated by a Namespace label calledrisk:sort_desc( sum( max(kube_namespace_labels{label_risk="$risk_query"}) by (label_risk, namespace) * on(namespace) group_right(label_risk) sum by (namespace) (avg_over_time(nam...
Include metrics of Namespaces where filtering label is undefined
Thedocumentationindeed indicates that the problem is Content-Type:The default behavior of a view function supports a request body of application/json. When a request is made with aContent-Typeof application/json, theapp.current_request.json_bodyattribute is automatically set for you. This value is the parsed JSON body....
I'm attempting to make a curl request to my python api that is using the AWS package Chalice.When I try to access theapp.current_request.json_bodya JSON Parse error is thrown. Cannot figure out why this is happening. My JSON is formatted properly as far as I can tell.Here is the curl request:(echo -n '{"data": "test"}'...
Cannot access the request json_body when using Chalice
TextBox.Text keeps the text as a simple string, it doesn't care about the real "meaning" of the string. Then, if you want to have your number back, you need to parse the string, hence neither implicit nor explicit cast to int is allowed (or better, it will throw an exception if you do it...). About the size, that tex...
I have a brief discussion with my teammate regarding this. He says, if I enter a number in textbox, and try to use the value later on using textbox.text or val(textbox.text), I will not need to parse the value to integer. According to him, if the text attribute value is all number, you can directly get the value as in...
Does textbox save an all number text as a long or string?
At the end I succeeded in using reference to the parameters (#p1, #2) during lookip and reference to#result.*during persist:@Cacheable(cacheNames = "PERSONS", key = "#p1 + #p2") Person findByFirstnameAndLastName(String firstname, String lastname); @CachePut(cacheNames = "PERSONS", key = "#result.firstName + #result.la...
How can I use@CachePuton an object, and update a cache bymultipleproperties of it?Example: Cached persons are added toPERSONScache every time the 'findOneByFirstnameAndLastname()` method is invoked.If thePersonobject is persisted, I also want the cache to update the person. But how can I tell@CachePutto use bothfirstna...
How to define multiple keys for @CachePut?
The only reason why this code not workingRewriteEngine On RewriteCond %{HTTPS} !on RewriteRule (.*) https://%{HTTP_HOST}%{REQUEST_URI}this is what the certificate give youcloudflareor something other similar service, when you hasn't direct control to your ssl certificate. I had such problem with ssl from cloudflare, an...
I have this .htaccess fileRewriteEngine On RewriteBase / RewriteCond %{HTTP_HOST} ^www\.(.*)$ [NC] RewriteRule ^(.*)$ http://%1/$1 [R=301,L]I want to modify.htaccess, that site always showhttpsin url. I triedRewriteEngine On RewriteBase / RewriteCond %{HTTP_HOST} ^www\.(.*)$ [NC] RewriteRule ^(.*)$ https://%1/$1 [R=301...
.htacces redirecting from http to https issue
I would recommend usingLambda@Edgeto append any headers that you're looking for to your origin response before it is returned to the viewer.It can be done as simply as the below example when added as a Origin Response event.import json def lambda_handler(event, context): response = event["Records"][0]["cf"]["re...
I test my website usinghttps://observatory.mozilla.org/analyzeand I got F score.The reasons are:Content Security Policy (CSP) header not implemented X-XSS-Protection header not implemented X-Frame-Options (XFO) header not implemented ...I serve my website using CloudFront.Where I put those missing headers to Cloud...
How to add headers to CloudFront response?
Naturally since you're keeping many thousands or tens of thousands of connections open you will be incurring a memory cost by doing so. Here's what I would ask: Are you adding things to an array or an object that is filling up? Are you instantiating heavy modules for each open connection? Can you run a profiler to see...
I am working on a project which need's to handle many ten's of thousands of active socket concurrent connections. Now been looking at the memory usage and the V8 engine, I am using Node.js v0.10.19 (Stable) I can't afford for "system out of memory" event to happen. So currently working on running up to 15 node instanc...
What is the best way to avoid "system out of memory" in node.js?
I forgot to add checkout action to workflow .yaml file... jobs: build: runs-on: ubuntu-latest steps: # Checks-out your repository under $GITHUB_WORKSPACE, so your job can access it - uses: actions/checkout@v3 - name: Install SSH Key uses: shimataro/ssh-key-action@v2 ....
I have added workflow to github actions where i want to rsync files from repository to remote serverjobs: build: runs-on: ubuntu-latest steps: - name: Deploy with rsync run: rsync -avuz --delete . ${{ secrets.USER }}@${{ secrets.HOST }}:~/App/MainPage/rsync sends nothing because my repository in...
Github repository is empty when rsync is used
Useps x | grep kubeletorcat /proc/$(pidof kubelet)/cmdlineto get commandline.If kubelet is installed byaptoryum, mostly it's working as a systemd service.Take a look at files in/etc/systemd/system/kubelet.service.d/folder, where the arguments kubelet running with.ShareFollowansweredNov 8, 2017 at 23:25silverfoxsilverfo...
Kubernetes kubelets can be run with a specific set of options (https://kubernetes.io/docs/admin/kubelet/). Is there a way to see, through kubectl or similar way, the options that kubelet was run with?I basically want to know if--allow-privilegedwas passed in, but see no way of checking that.
Check kubelet startup options/settings?
1 This is not trivial because docker doesn't track usage timestamps, only creation. So the builtin commands can't make decisions based on usage. An old but periodically active ticket to address this is https://github.com/moby/moby/issues/4237 I'm aware of two separate prog...
I know how to remove old Docker images created more that N days ago. See here But what I'd really like to do is to remove old Docker images that were not used for the last N days. The goal is to keep images that are frequently used, even if no container is actually using them when I do the cleanup. Is this possible?
How to remove old Docker images based on past usage