Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
SSH attemps to read the keys from$HOME/.ssh. Since you are running asroot, the script most likely fails to find anything under/home/root/.ssh.The fix? Trysudo -E, which preserves the environment variables. For more info, consultsudo(8).
I'm using Raspbian.I did successfully added the key and I can clone a private repo with ssh auth manually, using the commandgit clone[email protected]:USER/repo.git.Here is how I set it up:ssh-keygen -t rsa -b 4096 -C “email” eval "$(ssh-agent -s)" ssh-add -k ~/.ssh/id_rsa cat ~/.ssh/id_rsa.pubHowever, if I try to run ...
Git Clone ssh authentication through bash script
Looks like you hit a known problem in AKS clusters v1.19+ around "Pod IP SNAT/Masquerade behavior".How it affects clusters using Calico's plugin for Network Policies was explained there by other users:Just for information of other users, this issue causes problem for a NetworkPolicy with podSelector configs. Since the ...
I am using Azure Kubernetes Service (AKS) and want to make sure pods inside a specific namespace can only receive ingress traffic from other pods in the same namespace.I found this network policy to achieve this namespace isolation (fromhere):kind: NetworkPolicy apiVersion: networking.k8s.io/v1 metadata: namespace: m...
Kubernetes network policy blocks traffic between nodes on AKS
My understanding is, the line p[5000] = 3 should generate a page fault No, dereferencing p to a memory that was not allocated is undefined behavior: it does not imply that your program must segfault. For example, p[5000] may still access some memory your process owns and therefore no segfault occurs. The code works...
So I'm on x86-64 linux, and when I try to mmap one byte, I get way more than one page. Here's my thinking: When I allocate one byte, I should be able to access PAGE_SIZE bytes after that. That's how paging works, right? I confirmed that PAGE_SIZE on my system is 4096. But still, the following code doesn't segfault: ...
Mmap allocating more than one page
Created my own local hook that works well enough for my needs. Add this as a pre-push hook to your local clone of a repo: #!/bin/sh branch=$(git rev-parse --abbrev-ref HEAD) userRepo=$(git remote -v | grep fetch | awk '{print $2}' | grep "github.com" | cut -d':' -f2 | rev | cut -c5- | rev) if [ -n "$userRepo" ] the...
One of the things I find very handy about Bitbucket is when you push a new branch up to a repo hosted in Bitbucket, it prints out (to the terminal screen) a URL that you can hit to create a PR from that branch you just pushed. Ex: $ git push origin someBranch Compressing objects: 100% (3/3), done. Writing objects: 10...
Git hook to produce Github "Create Pull Request" link in terminal like Bitbucket Does
You are addingRewriteCondto wrongRewriteRule.RewriteCondonly in effect to very nextRewriteRule. Try this code:RewriteEngine On ## Unless directory, remove trailing slash RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^(.+)/$ $1 [R=302,L] RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d Rewrite...
I have a problem with mod rewrite apache seems to ignore the rewrite cond.The rewrite rule is working as all my pages are working but the problem is with resources like css, imgs and js. In my html I use"href=assets/css/style.css"which is an existing file but i i am redirected on my index.php...Is there something i did...
rewritecond not working for existing files
Regarding to (2): It's actually possible to recover them if you are quick enough. They should still be there, even when Google says they're deleted.If you know the name of the deleted DB run the following command to check if they are still theregcloud sql backups list --instance=deleted-db-name --project your-project-n...
I would have two questions related to cloud sql backups:Are backups removed together with instance or maybe they are left for some days?If no, is it possible to create new instance from backup of already gone instance?I would expect it possible but looks like backups are only listable under the specific instance and th...
Create instance from backup on google cloud sql
0 Don't forget to use unset() after while/foreach to cleanup no longer required (DB) results. Share Follow answered Aug 25, 2011 at 15:09 powtacpowtac 40.8k2828 gold badges116116 silver ba...
I have created a PHP script which scrapes some particular data from a website, looks for the relevant information and store this in a database. The script works fine on the local machine but when i run it on a live server using cron job, it stops after doing 10 database inserts saying "Fatal error: Allowed memory s...
PHP: Very high memory usage on production server while memory usage is low on local machine
Try this solution - essentially, it just turns off the rewrite engine in that specific directory:http://wordpress.org/support/topic/exclude-directories-from-wp-rewriteAdd a.htaccessfile to the directory that you want the Wordpress rewrite to ignore, with this code:RewriteEngine offOr, try changing the last line you hav...
Client added a WordPress install to their document root directory. Afterwards one non-WordPress directory is giving a 404.In a browser you can access/foo/facebook/index.phpwith no issues but you get the WordPress 404 page if you try to access/foo_admin/index.php. Wordpress is in / (not pictured here due to clutter).H...
404 due to WordPress mod_rewrite rule?
You can specify an AWS IAM user using the AWS key instead of Service as the Principal for a role policy document, including an AssumeRolePolicyDocument: "Principal": { "AWS": "arn:aws:iam::AWS-account-ID:user/user-name" } Refer to the Specifying a Principal section of the IAM Policy Elements Reference for full detail...
So I'm constructing a cf stack for a role in AWS and I don't know how to go about the AssumeRolePolicyDocument field when designing a role that is not resource-based. All the examples I've tried to look up each have a specific AWS resource designated under the "Principal" field (e.g. "Service": "ec2.amazonaws.com"). ...
CloudFormation IAM Role -- AssumeRolePolicyDocument
I know this may be general but I decided I wanted to get this answered, so here is what I did and it is now working correctly: Uninstalled ALL git clients through windows (I had a couple initially for testing) Installed the basic git bash from git Turned off smart screen and ensured firewall and defender were off (th...
I am working from home and am trying to clone a repo from Github and am getting the following error: git clone https://github.com/account/repo.git Also tried git clone git://github.com/account/repo.git Error: Failed connect to github.com:443; No error while accessing https://github.com/account/repo/info/refs?service...
Git clone not working - error 443
I tried this solution and for me worksversion: '3.1' services: alp: image: alpine volumes: - my-jenkins-volume:/your/local/path command: sleep 10000 volumes: my-jenkins-volume: external: falseexternal true if you provide your volume from an external source, not directly from the doc...
I'm on docker version 1.11.2. I'm able to create named docker volumes:docker volume create --name my-jenkins-volumeThan I'm able to connect my container with the named-volume with the-voption:docker run -d -u jenkins --name jenkins -p 50000:50000 -p 443:8443 -v my-jenkins-volume:/var/jenkins_homeIs it possible to creat...
Create named docker volume with docker-compose?
Nvm, I fixed it.I ended up deleting my local repository and cloning it remotely from GitHub. That didn't solve it immediately, but the Gemfile on my local repository was from before I started having issues.After that, I then kept getting an error where the version of Jekyll was incompatible with building GitHub pages.T...
As the title above says, I'm trying to build a site on localhost withjekyll servebut I keep getting several errors.I've done all the steps describedhereto build my site, but nothing seems to work. I've tried several troubleshooting sources I found online, as well.Also, the directory I'm trying to build with is a local ...
Jekyll Serve command not working - Errors include MissingDependencyException and MissingSpecError
0 You can use any external plugins or apps to connect cpanel with dropbox or google drive . as a hosting company we know that few of our clients using https://backupcp.com for the personal purposes . Share Follow ...
I am using cPanel for managing my websites in Shared Hosting. Now want to store my backup on my Google Drive. I have check lots of tutorials all are said, we need to do something on WHM. I don't have that. CPanel only i have. Please help me to store cPanel backup files to Google Drive. Thanks.
How to do CPanel Backup to Google Drive
1 First of all please check if there is firewall. Disable it. Then remove the IP address provided in core.yaml file in operations tab. So that line will look like a below: operations: # host and port for the operations server listenAddress: :9443 This change worked...
I know this error has been posted before on StackOverflow and many solutions/answers are available too. But, I've already gone through all those answers and couldn't find any viable solution for me. I'm running a Hyperledger Fabric network with single peer and orderer. Their Docker containers have exposed the followin...
Getting error curl: (56) Recv failure: Connection reset by peer
I am not sure exactly how you would achieve it with docker or anything else, as i dont see anyway to ask tomcat to just expand the war before it actually starts. But as per standard practices its not a good idea to explode a war and tweak the contents. It kills the entire purpose of making a war.Rather you should make ...
I am using docker to deploy a tomcat container running a third partywarfile.MyDockerfilelooks something like thisFROM tomcat:7-jre8 ADD my.war ${CATALINA_HOME}/webapps/my.warWhen I run the container tomcat expands mywarat runtime and I can happily access my app athttp://my.ip.addr:8080/mywar/.However my problem is tha...
Docker tomcat edit expanded war files
No they are not included.ApproximateNumberOfMessagesis the number of messages currently available for retrieval:The number of messages available forretrievalfrom the queue.Also for Standard queue, these values are approximate only, and may not represent actual numbers of messages present:For standard queues, the result...
I'm currently in the process of using SQS from AWS. Every 60 seconds my application pulls all of the messages off the queue and processes them. Given that I can only retrieve up to 10 messages in a single request, I want to query the queue for theApproximateNumberOfMessages. The aproximate number of messages in the que...
Does the SQS ApproximateNumberOfMessages attribute include the delayed and not visible messages?
<div class="s-prose js-post-body" itemprop="text"> <p>why you want absolutly in SVG? If you reduce the PDF with Ghostscript, the images will stay like they are, just change resolution.</p> <p>I've convert PDF in smallers size with ghostscript. It's easiest and faster than with convert ( imagick).</p> <p>I think you can...
<div class="s-prose js-post-body" itemprop="text"> <p>Good day, I have read all questions and answers about "PDF to SVG", "PDF to PNG, PNG to SVG" and had no success in accomplishing my task.</p> <p>I have a PDF that is about 10mb (blueprints from archiCAD) and after using pdf2svg to convert to SVG, the SVG is about 70...
How to convert 10MB PDF to SVG quickly with appropriate size for web
It looks likeautolinked referencesare not supported in regular md files.ThisPosttalks about using absolute and relative references such as:[<hash>](https://github.com/user/project/commit/<hash>)[<hash>](/../../commit/<hash>)
In GitHub a commit SHA in an issue or comment is automatically hyperlinked to the commit when it renders on the browser, however in a markdown file (e.gChangelog.md) this does not happen (an example). Is there a way to get this same behaviour in markdown files?
How to get a commit SHA to link to Github commit in markdown file
Since no one answered it for 3 month, I'll put my own solution. Under /Shared/ dir in databricks we have notebooks which should be synced to repository under notebooks/Shared/. I run this script on regular basis, thus keeping all notebooks up-to-date in a repo. databricks workspace export_dir /Shared ./notebooks/Share...
I'm trying to set up GitHub integration for Databricks. We have hundreds of notebooks there, and it would be exhausting to add every notebook manually to the repo. Is there some way to automatically commit and push all notebooks from databricks to repository?
Databricks-GitHub integration, automatically add all notebooks to repository
Using ModSecurity, try this rule (last 2 lines are optional):SecRule REQUEST_METHOD "!@within GET POST PUT" \ "id:20,\ phase:1,\ deny,\ msg:'Method is not allowed by policy',\ logdata:'%{MATCHED_VAR}'"ShareFollowansweredMay 17, 2022 at 14:12azuritazurit11411 silver badge77 bronze badgesAdd a comment...
I want to create a rule that blocks all http requests (get,post,put, literally all of them) and only allow certain ones that I specify. Specifically, Get, Put and Post. I am running ModSecurity V3.If there is a rule that can do something like this, if someone wouldn't mind sharing the syntax, I would greatly appreciate...
How do I add a ModSecurity rule to block certain http requests?
You would need tosetup Git first with matlab.That involves:uncompressing thelatest git archiveanywhere you want (likePortableGit-2.7.0-64-bit.7z.exeuncompressed inc:\PortableGit-2.7.0-64-bit.addc:\PortableGit-2.7.0-64-bit;c:\PortableGit-2.7.0-64-bit\bin;c:\PortableGit-2.7.0-64-bit\usr\binto your%PATH%.define the enviro...
This is the first time I'm using Github. I want to use it as a tool for sharing code and organizing the different versions of my code.I've downloadedGit for Windowsand installed. I've also installedGithub Desktopas the GUI client. I'm studyingProGit bookandGitHub Desktop Documentationto start.Recently I'm coding with M...
How to set up Git source control for matlab?
From my deployment process, flannel is uncompatiblity with etcd v3.0 now(2019/09/06). I am finding a way to handle this, when I starting flannel it throws this error:E0908 20:05:41.432715 28891 main.go:382] Couldn't fetch network config: 100: Key not found (/kube-centos)ShareFollowansweredSep 8, 2019 at 12:12Dolphin...
I am planning to migrate our Kubernetes instance to etcd v3. Do you know if Flannel will work with etcd v3?
Flannel compatiblity with etcd v3.0
You could either push tasks from GAE towards AWS, or have your AWS instances pull tasks from GAE.If you push tasks from GAE towards AWS, you could useURLFetchto push your data towards your AWS instances.If you prefer to have your AWS instances pull tasks from GAE, you could have your GAE instances put their tasks in th...
I am currently using google app engine as my mobile application back end. I have a few tasks that can not be performed in the gae environment (mainly image recognition using opencv). My intention is to retain gae and use AWS to perform these specific tasks.Is there a simple way to pass specific tasks from gae to AWS? E...
Using amazon web services as google app engine back end
You can have a new rule for that:RewriteEngine On RewriteBase /app/ RewriteRule ^([a-z]+)/([a-z]+)/?([0-9]*)$ index.php?controller=$1&action=$2&id=$3 [NC,L,QSA] # If the request is not for a valid directory RewriteCond %{REQUEST_FILENAME} !-d # If the request is not for a valid file RewriteCond %{REQUEST_FILENAME} !-...
I have these pretty urls:app/category/overviewapp/category/newapp/category/edit/4app/product/overviewapp/product/newapp/product/edit/4.htaccessRewriteEngine On RewriteRule ^([a-z]+)/([a-z]+)/?([0-9]*)$ index.phpcontroller=$1&action=$2&id=$3 [NC,L]Is it possible to define that any other other url goes back to the index?...
Is it possible to allow only the correct pretty url?
AWS now offer a solution called Elastic File System (Amazon EFS) that lets multiple instances access a shared file store.
How can i attach persistent storage on Elastic Beanstalk ?I know i need to have a .config file where i set the parameters of the environment to run every time an instance is created.My goal is to have a volume, let's say 100GB, that even if the instances got deleted/terminated, i have this volume with persistent data w...
Persistent storage on Elastic Beanstalk
For those asking the same question and upvoting.This is my current sollution:For some time I treatedkey-file.jsonas an identity token, put it to the CI config and used it within container withgcloudCLI installed. I used the key file/token to log in to GCP and letgcloudgeneratekubectlconfig - the same approach used for ...
I deploy apps toKubernetesrunning onGoogle Cloudfrom CI. CI makes use ofkubectlconfig which contains auth information (either in directly CVS or templated from the env vars during build)CI has seperateGoogle Cloudservice account and I generatekubectlconfig viagcloud auth activate-service-account --key-file=key-file.jso...
"permanent" GKE kubectl service account authentication
I share this link.Install using native package managementUpdate the apt package index and install packages needed to use the Kubernetes apt repository:sudo apt-get update sudo apt-get install -y ca-certificates curlDownload the Google Cloud public signing key:sudo curl -fsSLo /usr/share/keyrings/kubernetes-archive-keyr...
I am trying to set up Kubernetes in my instance on a ubuntu 22.04 virtual machine. While in the installation process i am trying to add a gpg key for intra cluster communication. After running the command:sudo curl -s https://packages.cloud.google.com/apt/doc/apt-key.gpg | sudo apt-key addI got a message thatapt-keyadd...
What commands (exactly) should replace the deprecated apt-key?
Ok, so the issue is that you have to guess that the package you need to install is called "libzip-dev", not just "libzip".So the solution was :RUN apt-get update \ && apt-get install -y libzip-dev \ && docker-php-ext-install zip
I want to install php-zip on my docker image (the end goal is to use the PhpWord Library). I use the php:7.4-fpm, which runs on Debian.In my dockerfile, i use the command :RUN apt-get update docker-php-ext-install zipWhen executed, the script crashes because it can't find /usr/src/php/ext/libzip.So i add a good old "ap...
Installing PHP-zip on a php:7.4-fpm image
There is a difference between having the key incorporate the whole cert-chain, or having the key plus one or more certificates in your keystore.You may use openssl to create a key that incorporates the whole cert-chain, e.g.:openssl pkcs12 -in <key in .pem-format> -export -out tmp-keystore.p12 -certfile <one file with ...
I have Mountain Lion Server and an SSL certificate signed by Go Daddy. I have various services using the certificate successfully which is verified using several online SSL checker tools.I have recently set up Jira which runs under Tomcat so I need to use those same certificates imported into the keystore.I noticed tha...
Keytool unable to chain imported PCKS12 certificate?
Your memory is probably being fragmented from numerous string operations or other operations that create and release small blocks of memory, such as boxing/unboxing. You will get this exception when the CLR cannot allocate a large enough free block of memory. I use the "CLR Profiler" and check the memory allocations. ...
Inside a complex multithreaded application I am hitting an out-of-memory exceptions maybe once a week. The application is sending/reading massive amounts of data via several sockets, where the read data gets cached to avoid network card buffer overruns. What is the best strategy to analyse the memory exceptions? Durin...
C# Out of memory exception - warning strategy
The Gunicorn documentation suggests that you can specify SCRIPT_NAME through an HTTP header without any further explanation. Digging through the source code revealed that it accepts a non-standard header actually named SCRIPT_NAME. The following can be used to set SCRIPT_NAME for Gunicorn: proxy_set_header SCRIPT_NAME...
I'm testing Gunicorn as an alternative to uWSGI. An Nginx reverse-proxy is set up in front of the Gunicorn instance. A consequence of not using uWSGI is I have to resort to using Nginx's proxy_pass instead of uwsgi_pass. With uwsgi_pass I could use uwsgi_param to overwrite the PATH_INFO and SCRIPT_NAME WSGI variables....
Nginx and Gunicorn WSGI Variables
The problem here is that you have allocated just memory, yet you have declared your object as a set of pointers to pointers. The double subscript requires a pointer vector pointing to each row or a type that tells the compiler how many columns there are. So some options are: do the subscript calculation with code, pe...
I'm using malloc to allocate memory for a matrix, but afterwards any access I do to the matrix to set an element for example gets me a segmentation fault. This is what I'm doing right now: int **matrix = malloc(rows_number * columns_number * sizeof(int)); if (matrix) matrix[0][0] = 1; Why can't I access the matrix a...
Allocating memory for a matrix in C, why can't I access the matrix afterwards?
1 How to ensure confirmed graceful exit of docker? Not by looking at the docker container ls -al output, which as commented shows the stopped containers. But by looking at: the logs of a stopped container, for an application message stating the stop was graceful (so that...
Docker sometimes exists but still keeps running. Started with : docker run -v $PWD:/host --rm -it Ubuntu_2018 After "exit" : Docker container ls; docker container ls -al This still shows container running How to make sure that docker container is gracefully ended? Using the option "--rm" and exiting the docker with "...
How to ensure confirmed graceful exit of docker?
From your question: I want to avoid getting the browser popup that asks the user to 'resend.' I want the browser to just use the copy of the page it has in it's cache. If browser asks you to resend data, it means that content was response to POST request. According to RFC 2616 - Hypertext Transfer Protocol -- HTTP/1...
I added an outputcache directive to my asp.net page (asp.net 4.0) as such: <%@ OutputCache Duration="3600" Location="Client" VaryByParam="None" %> However, this does not appear to be working. When I check the http header information I see this: HTTP/1.1 200 OK => Cache-Control => no-cache, n...
OutputCache directive not working in Asp.Net
Lets keep aside about searching a repositories using API's for moment to understanding the expected results When you search the C++ from github search engine then you will find very less amount of C++ repositories were C repositories also fetched. But when you search for cpp from github search engine then you will get...
I'm using this api: https://api.github.com/search/repositories?sort=stars&order=desc&q=language:C++ to find C++ repositories. But the return page is all language:C, Why?
github search api can't find language:C++
I found the problem finally. My C# project have some code to detect its memory usage, when it occupy memory over 1GB, it will throw OutOfMemoryException itself. After I comment these code, the program can reach memory usage to 3GB.
I take over a C# project which loads 3D models into memory, so I need large memory to use. My platform is 64-bit win10, the C# program is 32-bit, and I use visual studio 2013 to develop. My laptop have 8GB memory. Before I use editbin /largeaddressaware $(TargetPath) to add LARGE_ADDRESS_AWARE flag to the C# program...
Except LARGE_ADDRESS_AWARE, what else factors will limit C# process memory consumption?
Yes, the standard memory in your link is global memory. There's a program deviceQuery in CUDA samples, which shows the info you need. You could find it in CUDA dir, typically in /usr/local/cuda/samples/1_Utilities/deviceQuery. The following document also tells you how to build it. http://docs.nvidia.com/cuda/cuda-sam...
I want to know what is the shared memory and global memory size of my GPU. I am using nvidia tesla k40c. I found this link and it says standard memory is 12 GB. Is standard memory the same thing as global memory? How can I find size of shared memory? is there any command that shows gpu features? I used nvidia-smi comm...
How to find out shared memory and global memory size of GPU?
You could passgrafana.envvalueto addSMTP-related settings:GF_SMTP_ENABLED=true,GF_SMTP_HOST,GF_SMTP_USER and GF_SMTP_PASSWORDshould do the trick. Theprometheus-operatorchart relies on the upstreamstable/grafanachart (although,still using the 1.25 version)
I used HELM to install the Prometheus operator and kube-prometheus into my kubernetes cluster using the following commands:helm install coreos/prometheus-operator --name prometheus-operator --namespace monitoring --set rbacEnable=false helm install coreos/kube-prometheus --name kube-prometheus --set global.rbacEnable=f...
Updating Grafana configuration file to configure custom SMTP settings
Let's assume you're working onmy_branch, and made some changes there. Now your co-worker pushes to the master, you need to update yours:git stash git checkout master git pullYourmasteris synced with the changes.Now you want to continue working onmy_branch, so yougit checkout my_branchand now you shouldstash popthe chan...
I'm having trouble with syncing my local repo with the remote repo on GitHub if both remote and local repos have changes. So, for example, if I haven't started working on a feature, before I start I can sync by simply runninggit pull origin masterThis is a simple one. However, while I am working on a feature on local, ...
Stash / Commit before Pull Not to Lose Local Files
Here's the answer to my question:CTRL+Ain the GIT GUI is actually to ADD A NEW REMOTE and is not the same as GIT ADD ! In the GIT GUI, there does not seem to be any distinction between 'adding' and 'staging' , both are done by the 'stage to Commit' command:CTRL+T. And according to this:http://hoth.entp.com/2008/11/5/wh...
If git addCTRL+Astarts tracking changes AND stages changes to already tracked files, why should I ever use Stage to CommitCTRL+Tin the git gui ?
What's the difference between Git add (ctrl+A) and Stage to Commit (ctrl+T) in the git GUI?
Use a query that returns all the values for which you want results and left join the table to aggregate:select t.reference, count(l.id) from ( select 'ref1' reference union all select 'ref2' union all select 'ref3' ) t left join db.log l on l.reference = t.reference group by t.referenceSee a simplifieddemo.
I have a table calledlogwhich contains logs sent by several applications. This table has avarcharfield calledreference.I have a table panel in Grafana in which I show how many logs we have grouped byreferencevalues. So the user types one or multiple values in a text field on Grafana like'ref1', 'ref2', 'ref3'and a quer...
Include rows which don't match with IN() clause
Edge don't cause the security error, but GitHub's email address verify.Repro:GitHub default email address is E1,push code with email address E2, then cannot click the button.Resolve:GitHub=> Settings=> Emails=> add email address(E2)
I access my GitHub repo with Microsoft Edge. But when I click buttonClone or download, no response.With F12 Developer Tools, I get this when refresh page:SCRIPT5022: SecurityError frameworks-b656af906df0ec252f7683ae76db8ed358bf5b309bd5781fc467ae3c1d571c2a.js (1,6939) SCRIPT5022: timeout loading module: github-bootstrap...
How handle 'SCRIPT5022: SecurityError' when access GitHub with Microsoft Edge
Unfortunately yes. You have to specify thekeypairwhen you launch the instanceunlessthe AMI comes with a well known username/password and the SSH daemon accepts both password and key.In fact, AWS warns you when you proceed to launch an instance without akeypair.
I'm new to AWS.I created an instance on EC2. It's a Bitnami Wordpress install.I didn't realize that I needed a key pair, and when I added one after the fact, it does not allow me to connect. I get "Permission denied (publickey)."Can I disconnect the instance without losing the Bitnami WP install?Will doing so allow me ...
AWS: add key pair after creating instance
Kubernetes allows you to create a service without selector, and cluster will not create related endpoint for this service, this feature is usually used to proxy a legacy component or an outside component.Create a service without selectorapiVersion: v1 kind: Service metadata: name: my-service spec: ports: ...
I have minikube version v0.17.1 running on my machine. I want to simulate the environment I will have in AWS, where my MySQL instance will be outside of my Kubernetes cluster.Basically, how can I expose my local MySQL instance running on my machine to the Kubernetes cluster running via minikube?
Minikube expose MySQL running on localhost as service
10 I don't think that you can do that directly from the GUI of Rstudio. You can (in Rstudio) go to GIT -> More -> Shell and perform the commands, but this is similar to just working from git bash. (fatherly advice: learn git from the command line, it will benefit you in t...
I am a beginner to Git and GitHub. I came across tutorials on working with Git using RStudio. Instead of working from git bash RStudio offers a GUI type connection with Git and Github. Can someone tell me whether is it possible to merge git branches in RStudio? If so, how to do that? Thank you :)
How to merge git branches in RStudio
Currently we havetwo optionsto solve this:Option 1:Duplicated locations: NGINX looks for the best match. (a little better performance)location /post/ { post config stuff; . . . } location ~* ^/post/.*\.(css|js|png|gif)$ { post/files.(css|js|png|gif) config stuff; expires max; add_header ...
I have pictures, and I want to add their headers to max, I have profile pictures which can be changed and post pictures, I want to add headers only for post pictures, but not to profile pictures, I have no idea how can I manage this. thank you, this is my configuration,this is the path of posts, /post/name-of-the-pictu...
How to add headers to only specific files with nginx
The form of a cronjob is:* * * * * user-name command to be executedYou forgot to add the user which runs the cronjob. Supposing you run it using the root user, change the code to:*/2 * * * * root /usr/bin/python /usr/lib/python3.5/venv/send_email_attachment.py > /tmp/listener.log >/dev/null 2>&1
I read some similar topics but no luck to fix my issue.This is my first Python script, it runs no problem when I manually execute it.But it doesn't work with below crontab.0,15,30,45 * * * * /usr/bin/python /usr/lib/python3.5/venv/send_email_attachment.py >/dev/null 2>&1I also try to below.*/2 * * * * /usr/bin/python /...
Python script doesn't execute via crontab
You can use the package dbms_metadata.get_ddl to generate ddl for the objects. Read this You can generate csv files for the data that you can load with sql*loader, or generate insert statements. This will be a real pain. Especially when there is a tool available that does it for you : exp/expdp. You could try to use a...
I have an Oracle account (schema) of a remote Oracle database. By using this account I can do all kinds of operations such as: query, update, delete and add. But I can't use exp or imp. Now my question is how can I dump all the data in this remote database schema, include table, view, function, procedure, package.
How can I backup a remote database schema in this environment?
Tarfile.next() method caches its contents, including the lines: if tarinfo is not None: self.members.append(tarinfo) It turns out that Tarfile.getnames() calls Tarfile.getmembers() which calls Tarfile._load() which repeatedly calls Tarfile.next() until all is read into self.members. So Tarfile.getnames() and it...
I'm trying to create a list of file names from a large tarball file and I would like to understand why the memory usage is still the same in my example? Is it because f.write() is still holding/buffering all objects in memory before the file is actually closed? Is there a way to improve this? # touch file{1..100000}.t...
Python: memory usage while writing to a file (generator vs list)
As stated by @user7294900, Sonarqube analysis requires JDK 11 for performing code analysis.Solution:You can either migrate your codebase to version Java 11, or use multi JDK build. Build your code with JDK 7, and perform sonarqube task with JDK 11.Steps for this is mentioned here,https://docs.sonarqube.org/latest/analy...
We haveANTbased project running on Java1.7version, while integrating Sonar in this project, we are getting below error on "sonar : sonar" (Sonar scanner analysis) line in build.xmlJava Version : 1.7ant version : 1.9sonar-ant-task-2.2.jarSonarQube Server : 7.5.0I know the reason of below error, but do we have any sonars...
UnsupportedClassVersionError: org/sonar/api/utils/SonarException : Unsupported major.minor version 52.0
In other words, the traffic will go on the public internet or will stay inside the VPC?It will go tointerface endpoint. And the reason is that aws choosesmore specific(longest prefix match) route when there is more than one choice where to direct traffic. So if VPC has to decide between0.0.0.0/0for NAT and interface en...
Say I have an ec2 in a private subnet with access to the internet through a NAT Gateway and I haveVPC endpoints (PrivateLink)for AWS services like DynamoDB or SNS. If my ec2 instance needs to interact with those services, it will use the private link or the NAT Gateway? In other words, the traffic will go on the public...
AWS private subnet with NAT gateway and VPC PrivateLink: which one will be used?
If the files is already in version control it will stay in version control, meaning other users will see it. You will need to remove the file manually. To delete the file but not remove it from your directory you can usegit rm --cached fileTo do it to every file you can usegit rm -r --cached .The file/files will not ...
What happens if you have a file on remote, that is locally on .gitignore, when you do a pull? Will you overwrite the content of the local file?
Git - File is on github but also on .gitignore - What happens on git pull?
If you installed them with pip, you can usepip freezeto list the currently installed modules. Save this to a file and usepip install -r fileon a new system to install the modules from the file.
Is there a way to backup Python modules? I installed lots of modules. If my system does not work properly, I will lose them all. Is there a way to do this?
Is there a way to backup Python modules?
You need one more rule to handle two parameters:RewriteEngine On RewriteCond %{HTTP_HOST} ^www\.(someserver\.com)$ [NC] RewriteRule ^ http://%1%{REQUEST_URI} [R=301,L,NE] RewriteRule ^(\d+)/?$ index.php?comp=$1 [L,QSA] RewriteRule ^(\d+)-(.+)/?$ index.php?comp=$1&string=$2 [L,QSA]
im quite new to rewrite rules. I can manage with one variable and thats it.I have webpage Where the rewriterule is:RewriteCond %{HTTP_HOST} ^www\.someserver\.com$ RewriteRule ^/?$ "http\:\/\/someserver\.com\/" [R=301,L] RewriteRule ^(\d+)*$ ./index.php?comp=$1 RewriteRule ^(\d+)*/$ ./index.php?comp=$1And it all work ...
ReWrite rule not working with two parametrs
-2Why not put every thing in a POST request?like this<form method='post' action='/'> <input type='text' name='data' value='some value' /> </form>the result in _request will bearray( 'data' => 'some value' )and you still retain your url structure.ShareFollowansweredJun 4, 2011 at 21:29Mient-jan StellingMient-jan St...
This question already has answers here:Closed12 years ago.Possible Duplicate:What is the best way to implement a friendly URL that has multiple variables using mod_rewrite?I'm usingmod_rewritein my site for friendly URLs which works great. When using a form to add a variable to$_REQUESTit's being called in the regular ...
HTML form directing to a friendly URL [duplicate]
You can pass config while creating service objects. Following is what you are looking forconst s3 = new aws.S3({ /* s3 config */ }); const ses = new aws.SES({ /* ses config */ });
I need to use multiple AWS credentials for different services like s3, SNS....var awsS3 = require('aws-sdk'); var awsSes = require('aws-sdk'); awsS3.config.update({ region: config.awsRegion, accessKeyId: config.sesAccessKeyId, secretAccessKey: config.sesSecretAccessKey }); awsSes.config.update({ regio...
How to set multiple aws credentials in nodejs aws-sdk module?
Entirely correct.sdk.class.phpis a file that exists in SDK 1.x, but not 2.x.The correct instructions are in theSDK2 README.
Environment: MAC - Mountain Lion I am trying to use the AWS PHP SDK for a project. I followed the Amazon web site's SDK installation directions (through composer) -- using the followingLink to AWSI created the file compser.json. Contens:{ "require": { "aws/aws-sdk-php": "2.*" } }From the command line, I typed...
AWS PHP SDK with Composer - missing sdk.class.php
As I understand your question, you want to set up your dedicated number of CPU for each app/pod. As I've searched.I am only able to find somedocumentationthat might help. The other one is a Github topic I think this is aworkaroundto your problem.This is a disclaimer, based from what I've read, searched and understand t...
I need to restrict an app/deployment to run on specific cpus only (say 0-3 or just 1 or 2 etc.) I found out about CPU Manager and tried implement it withstaticpolicy but not able to achieve what I intend to.I tried the following so far:Enabled cpu managerstaticpolicy on kubelet and verified that it is enabledReserved t...
Pin Kubernetes pods/deployments/replica sets/daemon sets to run on specific cpu only
Try to remove and redeploy the lambda.Also, make sure it has permissions to write to CloudWatch.
I accidentally deleted a lambda log group in CloudWatch.Now my lambda fails and I do not see the log group reappear in CloudWatch.Is it supposed to be recreated automatically? How can I fix the situation?I tried recreating the log group manually but it didn't receive any log.
AWS Lambda log group not recreated after deletion
Ideally when performing such a task like "replacing a node" below steps should be considered:Drain nodekubectl drain NODE_NAMEReset that nodekubeadm resetin the old node (optional step if the old node is accessible)Finallykubeadm delete node NODE_NAMEThings to consider when replacing a old node with new node:The new no...
Had to change node (server) with the new one leaving thesame node name. What I did was:master> kubectl delete no srv1 (removing old node)srv1> kubeadm join... (joining new node)after new node joined cluster no pods can be created.Warning FailedCreatePodSandBox 16s kubelet, srv1 Failed create pod sandbo...
Unable to run pods on new node
I had the same problem this morning and the following fixed it for me:$ docker-machine restart default # Restart the environment $ eval $(docker-machine env default) # Refresh your environment settingsIt appears that this is due to the Docker virtual machine getting itself into a strange state. There is an opengi...
I installedDocker-Toolboxjust now while following theirwebpageI started withDocker QuickStart Terminaland see following## . ## ## ## == ## ## ## ## ## === /"""""""""""""""""\___/ === ~~~ {~~ ~~~~ ~~~ ~~~~ ~~~ ~ / ===- ~~~ \______ o ...
Network timed out while trying to connect to https://index.docker.io
Use vectors!!! Since you know the number of elements, make sure that you reserve the memory first (by calling myVector.reserve(numObjects) before you then insert the elements.). By doing this, you will not call the default constructors of your class. So use std::vector<DataType> myVector; // does not reserve anything ...
I'm reading in values from a file which I will store in memory as I read them in. I've read on here that the correct way to handle memory location in C++ is to always use new/delete, but if I do: DataType* foo = new DataType[sizeof(DataType) * numDataTypes]; Then that's going to call the default constructor for each ...
C++ Allocate Memory Without Activating Constructors
You have two options:Forget forks.Use branches. This is the simpler option. Everyone involved needs access to the repo.Keep forks,use Pull Requests. This allows the owner of the repo to fully control whether changes are accepted or rejected.
I have a Github repository, lets call it Repo A. I have two other users than need to push/pull changes from this repository. How do I go about doing this? I have tried having one of the other users fork Repo A into their own Repo, lets call it Repo B, but when they make changes to the contents, it only pushes to Repo B...
How to push and pull to repository from forked repository
I think you should query likeAudited::Adapters::ActiveRecord::Audit.where("created_at >= ?", Date.today)according to thegem structure
I have rails 3 application with some models, like Product and User. I'm using "audited" gem to track changes for products, it's simple and nice working.But I want to make special page where I want to put daily activity history. I need something like Audits.all.order("created_at") for first step, but there is no such mo...
Activity history with Rails audited gem
Note that the pattern matching onRewriteCondislexiographic, notnumerical.111 < 12 3523 < 34 etc.
So this should be pretty simple, but it isn't working. I have the following code:RewriteEngine on RewriteCond %{TIME_DAY} >12 RewriteRule /orderforms/page1.aspx.*$ https://www.site.net/orderforms/page2.aspx [R=301,QSA]But even though it's not the 12th yet, it's still redirecting. Any ideas? Is this not fully supporte...
htaccess redirect based on date
2 If that background image is loaded as a Drawable from a Resource, it's bitmap data is already shared between all Drawables created that way. So no, it probably wouldn't benefit you in your particular case. Share Improve this answer F...
We have an Android application that throws out of memory error on Android 1.6. It is working fine on 2.0 onwards. We are trying to cut down on the runtime memory usage at this point. We were thinking of reusing images – such as the background image of each screen. Is it possible to load the background image once in t...
Android App - Can we Reuse Images in the Run time?
[RESOLVED] the problem was in network of docker, then i executed the folow command: docker build -t cassiamani/nodeapp --network=host . --network=host use dns from host, for me it's worked. reference: https://www.linode.com/community/questions/19710/docker-npm-timeout-on-linode
I'm try build image to node app, but when i execute docker build showed an error left run npm install npm ERR! network timeout at: https://registry.npmjs.org/cookie/-/cookie-0.4.0.tgz follow docker file content: FROM node:13 # Create app directory WORKDIR /usr/src/app # Install app dependencies # A wildcard is used ...
Docker build npm install error network timeout
According tothisyou need to use regex.path: /foo/.*
We have various Asp.net core containers with REST APIs hosted in Azure Kubernetes Service. Each REST API service implements many HTTP methods with different routes but has common base route (based on controller name).Requirements is to route the request to different service based starting route value, e.g. All requests...
Kubernetes ingress routing- Asterisk (*) is not working in ingress Path rules
Well I came up with this solution similar to the PDO one. Are there any unforeseen problems with running this as a cron job?<?php $con = mysql_connect("localhost","root","123456"); $throttle = 0; $batch = 50; $pause = 10; // seconds if (!$con) { die('Could not connect: ' . mysql_error()); } mysql_select_db("mai...
Let me rephrase my question, I have a mysql database that is holding emails to be sent, on a shared host. I would like to run a cron job that will read the database and sent out any messages in the database every 10 minutes or so.Now my question is, what is the best way with php to read my database and send out the ema...
PHP, Email and Cron
Make sure you copy past the SSH key exactly like it's in the xxxx.pub file do not add lines or spaces to it.Try again or delete both public and private keys and generate them again. It should work. Also remember to update the pub key again on Github website.Also test your existing key with this commandssh -T[email prot...
I installed Git and ran the set up my username and email.git config --global user.name "Your Name" git config --global user.email[email protected]Add your public keyNext, I created a project directory called 'projectname' and committed this to a remote Git server.mkdir projectname cd projectname git init touch README g...
The remote end hung up unexpectedly
Create a service and deployment type Job for master and slave. Job kind in k8s take care of deleting the pods upon completion.
I'm running Locust in master-slave mode on k8s, I need to auto delete all locust pods when a test is completed. Any suggestion? Thank you.
Locust master-slave on k8s, delete all pods when test complete
I found the resolution in anon-related question.To make SSL work without502 ERROR The request could not be satisfiederror, I had to specify my domain (e.g.example.com) as Alternate Domain Name in CloudFront distribution, attach the SSL certificate provisioned forexample.comand, very important, whitelistHostheader in th...
I need the end-to-end encryption, so current the setup is:example.com -> Route53 -> LB (TCP :80/443) -> Nginx Ingress with SSL certs installedIs it possible to do something like:example.com -> Route53 -> CloudFront -> LB (TCP :80/443) -> Nginx Ingress with SSL certs installedWhen I'm attaching the same certificate to C...
AWS CloudFront to EC2 end-to-end encryption
You are probably not using theNEflag.By default, special characters, such as&and?, for example, will be converted to their hexcode equivalent. Using the[NE]flagprevents that from happening.
I'm trying to redirecthttp://terrychew.com.sg/classestohttp://terrychew.com.sg/math-olympiad-singapore#Primarybecause without the#Primary, the tab I want users to see will not be selected..htaccess converts my#into%23which doesn't work for me. I read on another thread that this code can be used to redirect with hashtag...
how to rewrite url with hash # using htaccess
If youset your Jenkins url in the webhooks and services of your various GitHub repos, Jenkins will be called by that webhook.But theJenkins GitHub pluginwill only trigger the job where "Github project" has the right[email protected]:Person/Project.gitrepo.One possible workaround would be to:publish on GitHub one parent...
I have one standard Jenkins job that needs to be triggered for any pull request/commit made to any of the x repository I've added to the multiple SCM plugins in the configuration.Is there any way to have a hook on all the repos so that on a pull request to any of the repos the job builds the specific repo where the pul...
track multiple github repo on commit/pull request to build a jenkins job
Instead ofSHOW TABLEScommand, you could query theINFORMATION_SCHEMAdatabase. This way you could easily dump every table for every database and also know how many tables there are in a given database (i.e. for logging purposes). In my backup, I use the following query:SELECT DISTINCT CONVERT(`TABLE_SCHEMA` USING UTF8)...
I have done some digging around and I can not find a way to make mysqldump create a file per table. I have about 100 tables (and growing) that I would like to be dumped into separate files without having to write a new mysqldump line for each table I have.E.g. instead of my_huge_database_file.sql which contains all the...
mysqldump table per *.sql file batch script
I wasn't able to solve this via the sqljdbc drivers so I took some advise from another post and fell back to the jTDS driverhttp://jtds.sourceforge.net/index.htmlThe post I used wasSonar MsSql Database IssueMy Setup:Sonar 2.11 unzipped to C:\Sonar\sonar-2.11The steps I made (paraphrased)Download jTDS drivers fromjTDS D...
When I start Sonar (StartSonar.bat), I get the following error in the log file.Wrong column type in SONAR.dbo.rules for column description.Found: ntext, expected: nvarchar(max)Looking at the column in SQL Server Management Studio it isnvarchar(max)I'm running:DB: SQL Server 2005OS: Windows 7 64 bitSonar: 2.11Tried thes...
Error Running Sonar connected to SQL Server 2005. SONAR.dbo.rules for column description
I always prefer to use (as documented here): git config --global user.useConfigOnly true That way, I am forced to set the right user.name/email on each new local Git repository, in the local Git configuration of the specific repository. Any tool manipulating the repository (like VSCode making new commits) would then ...
I ran into a problem where I have made edits to a repository(org repo) and committed them with my personal git account accidentally. As a result, the commit history on the repo shows that the Committed & Authored identities are from the personal account rather than the official account. Are there any plugins or valida...
Configuring Git Account VSCode
RewriteEngine On RewriteRule (.*)/([0-9]{8})$ /$1 [R=301,CO=cookieName:$2:.example.com:1440:/]take a look athttp://httpd.apache.org/docs/current/rewrite/flags.html#flag_cofor more information about cookie flag
I am completely lost and all my searching on Google and winthin the board did not get me to a working solution.SetupApache-webserver htdocs are located: xampp/htdocs my Joomla site is located: xampp/htdocs/example.com/wwwTwo kinds of URI as possible on our homepage:URI end with a number (8 digits; e.g.: www.example.com...
htaccess - how to cut tailing number for URI and save it in a Cookie
You use try horovod with PyTorch.ResNet50 example is here:https://github.com/horovod/horovod/blob/master/examples/pytorch/pytorch_imagenet_resnet50.pyhorovod-related changes should be small and isolated.
I have a cuda9-docker with tensorflow and pytorch installed, I am doing cross validation on an image dataset. Currently I am using a for loop to do the cross validation. Something likefor data_train, data_test in sklearn.kfold(5, all_data): train(data_train) test(data_test)But the for loop takes too long, will the ...
pytorch parallelize for loop of Cross Validation
Yes, you can do this using thegit push -fcommand. The-fcommand is the "force" option, which pushes your current ref to the upstream, even if your current history is not derived from the upstream.You can easily lose work with this command.Be careful.
I have code on my client that conflicts with a checked in version. Is there a way to tell Git to "ignore the origin version and force my version over it"?
GitHub command to force a checkin
The key is:error removing unix socket, unlink(): Permission denied [core/socket.c line 198]You (very probably) previously run a uwsgi instance as root creating the unix socket file with root permissions.Now your instance (running instead as www) is not able to re-bind() that socket as it is not able to unlink it (no p...
Could not start uwsgi process via ini flaguwsgi --ini file.iniNot any uwsgi pidsps aux | grep uwsgi root 31605 0.0 0.3 5732 768 pts/0 S+ 06:46 0:00 grep uwsgifile.ini[uwsgi] chdir =/var/www/lvpp/site wsgi-file =/var/www/lvpp/lvpp.wsgi master = true processes = 1 chmod-socket=664 socket = /var/www...
Could not start uwsgi process
You need to ensure your view has: {!! csrf_field() !!} and you don't need to have it as a hidden field in the form.<form class="form-horizontal" method="post"> @foreach ($errors->all() as $error) <p class="alert alert-danger">{{ $error }}</p> @endforeach {!! csrf_field() !!} <fieldset>
When it comes to access my login I have this error:{"error":"token_not_provided"}I have in my .htaccess:RewriteEngine On RewriteCond %{HTTP:Authorization} . RewriteRule .* - [E=HTTP_AUTHORIZATION:%{HTTP:Authorization}]I have in my routes.php:Route::post('/api/login', 'AuthenticateController@authenticate'); Route::get('...
Login fail in laravel 5 (JTW)
There is nothing wrong with caching your database queries. For every category name you can store the ID in a dictionary as an example that expire after a certain period of time. This will remove the DB calls in this case. As an example: Dictionary<string, int> categoryIdLookup; This can be stored in the HTTP cache...
I'm performing a UrlRewrite for my main category pages. Converting: www.mysite.com/Category.aspx?id=2 to www.mysite.com/Dogs In order to do so I'm using Global.asax's Application_BeginRequest where I perform the following code(pseudocode): protected void Application_BeginRequest(Object sender, EventArgs e) { ...
UrlRewriting on Global.asax and SQL Output Caching
The simplest solution is to set credential.useHttpPath to True in your global Git configuration: git config --global credential.useHttpPath true According to the manual for git credentials: useHttpPath By default, Git does not consider the "path" component of an http URL to be worth matching via external helpers. Th...
Not even sure if I asked the question right. I have a repo for company1.com/foo which lives in directory ~/go/foo, for company foo. I have a repo for company2.com/bar which lives in directory ~/go/bar. I am hoping that when I'm working in the local repos git can look in the .git directory for completely different sets...
What is the term for managing GitHub permissions per directory?
Now that I have spent time on this, I am answering this question so that it can help someone. The answer is YES. Minikube provides a k8s setup which supports the OIDC based authentication. I have been able to configure it. So here is some details on how I configured the kube-apiserver parameters.minikube start \ ...
I have installed K8S using minikube on ubuntu 16.04 machine with VirtualBox driver.I am confused with various documents that are related to this topic. Some say it is not possible with minikube, but minikube documents that it suitable for test purpose. So i believe that maybe there is a way to achieve oidc authenticati...
Can minikube handle oidc authentication?
It makes sense. However, make sure you take care of possible race conditions. You could use the ReaderWriterLockSlim class to control access to the object from different threads. UPDATE: You absolutely need to lock the object appropriately, as the Add method is not an atomic operation. You can even leave the object in...
I want to create a "temporary cache lookup" to speed up a file lookup on my webserver. What I have now is a folder full of images, and if a user requests an image I use a File.Exists(...) to check to see if the image exists. If it doesn't, download it from another server, and either way, redirect the user to it. The ...
Is it advisable to use HashSet as a static global on a webserver?
I solved my problem, the solution is firstly (in my case) delete the three .htaccess files and edit this line in app/config/core.php//Configure::write('App.baseUrl', env('SCRIPT_NAME'));to this:Configure::write('App.baseUrl', env('SCRIPT_NAME'));it works for me, thanks guys anyway.
I just uploaded my website to the server but is not loading the css and js files, only I can see the site just like text and after put the .htaccess files in their places I got this message in the site:Internal Server ErrorThe server encountered an internal error or misconfiguration and was unable to complete your requ...
CakePHP 2.6.0 is not loading CSS and JS files
Authenticationis part of the Enterprise package. With that, you'll have dedicated support.
I have a question about authentication and security in Hazelcast Client. Which methods can I use to authenticate Hazelcast Clients? I'm using an open-source 5.0 version on k8s.Thanks Dũng
How to authentication and security in Hazelcast client
One thing you can do is write a simple Windows Service, which will parse the file once received.look intoFileSystemWatcherto detect the file. this way you don't need to schedule any task.Now as for the PHP script you should be able to start a command line process and execute the php.
Assume that we have a device which sends data (a file) back to the computer whenever something unexpected happens. Let's say if it detects a movement, it will send a data file to the computer.The computer should notice and parse this file. Afterwards, it should execute a PHP file to do other stuff.Currently we are usin...
How can I call a php file with an event on Windows machine?
You can use AWS web console from Amazon to do the same things from Google Chrome. As of now there is nothing similar to Elastifox on Google Chrome.
Is there any extension for chrome which is similar to Hybridfox or Elasticfox?
Chrome Extension for Amazon Cloud
Solution: I was trying to install Velero v1.2.0 on a Kubernetes v1.10 cluster (that I cannot upgrade at the moment). It turns out the two are incompatible. By downgrading to Velero v1.1.0, I was able to successfully install, and this error message disappeared.
I'm having the following issue installing Valero via the Velero CLI to and AWS Kubernetes cluster.velero install \ --provider aws \ --bucket $BUCKET \ --secret-file ./credentials-velero \ --use-restic --wait \ --plugins velero/velero-plugin-for-aws:v1.0.0The install fails, with the result:An erro...
Velero won't install, "additionalProperties cannot be set to false"
I don't think you can do that on Github. And also think is okay like that, because you shouldn't be able to "create" a new branch on a project you don't necessarily have push access. Even if youdohave push access to the upstream project i don't see it as a clean workflow.
My workflow: I have a fork (origin) of a remote project (upstream). I have created a branch, topic/TICKET. I want to create a pull request of my origin branch to the upstream project in a same-named, brand new "topic/TICKET" branch. In other words, I don't want my pull request toward the master branch or any existing b...
GitHub Pull Request to a new Upstream Branch
Check this issue, might be the answer, it has to do with the shell GitKraken uses by default in Windows:https://github.com/typicode/husky/issues/243
I'm trying to commit any kind of change to a branch in my repository using GitKraken and get the following error every time:whats strange is if I rungit commit -m 'test'it commits without any errors. If I comment out the below line in my package.json then I can make commits in GitKraken no problem:"precommit": "yarn ch...
GitKraken pre-commit fails but works in command line commit
squid:S1451 "Copyright and license headers should be defined"has two parameters:headerFormatisRegularExpressionInheaderFormatyou define expected text. If you setisRegularExpressionto true, then the text will be treated as regular expression.If you use normal text, then you need to put full license, e.g.:/* * Maven Plu...
All of my SonarQube code analysis Java projects are failing on this rule.Each source file should start with a header stating file ownership and the license which must be used to distribute the application. This rule must be fed with the header text that is expected at the beginning of every file.There is one parameter:...
SonarQube Rules squid:S1451 Copyright and license headers should be defined
1 You can test this command for install directadmin via docker: docker pull hoolia/directadmin You don't needed to any more act. More info: https://hub.docker.com/r/hoolia/directadmin/ Share Follow ...
I have VPS hosting with Direct Admin. There is Debian Jesse. I can use ssh. Now I want to install discourse using a docker container. Should I do something more than I can read on tips in the manual? I mean. I don't know nothing about port. For example Direct Admin using:80 and 443 and 2222. ? Should I set a new "port...
Direct Admin + Docker
Do the Following to Back up and Restore your Mautic. Zip you Current Directory then Download Export your Maultic Database on Server B Creat Db Import your Current Db Upload Mautic Files to the domain folder you wanna use. change Db Connection in app/config/local.php These should fix the ish.
I am a beginner in IT and Web Development. I would like to create a backup of the Mautic installed on a hosting server “A” and restore it in another server “B”. How do I do that? If it’s possible to automate the backup and the restoration, please tell me how to proceed.
Backup and restoration of Mautic
5 It seems like the best solution I can find, is to just leave the HTTP cache layer alone, and use a separate cache layer: const { RESTDataSource } = require('apollo-datasource-rest'); import { PrefixingKeyValueCache } from 'apollo-server-caching'; class MoviesAPI extends ...
Using the simple "Movies API" example from the documentation. I added a ttl to the getMovie function, so that the result is cached for 10 minutes. How can I invalidate the cache in the updateMovie function? const { RESTDataSource } = require('apollo-datasource-rest'); class MoviesAPI extends RESTDataSource { async ...
How to invalidate cache in Apollo Server RESTDataSource
Events for the last 5 Minutes or N mins is currently an Open issue you can follow thisgithub linkfor more feature enhancements. You can also raise a comment under this link or else you can raise a new feature request . We can also use the below commands as of now :This command gives all the events with the time stamp....
Is there anocorkubectlcommand that prints out only events of the last 5 Minutes?Something like:oc get events --last="5-Minutes"
Openshift cmd only get events of last 5 Minutes
Answering myself:This is a dnspython bug. Solution:pip install dnspython==1.16.0ShareFollowansweredJul 19, 2020 at 23:57Michael A.Michael A.1,0991212 silver badges2222 bronze badges0Add a comment|
I am running Celery in Kubernetes pod. It can't find the server:ERROR/MainProcess] consumer: Cannot connect to redis://:**@redis-master:6379/1: Error -3 connecting to redis-master:6379. Lookup timed out.. Trying again in 4.00 seconds... (1/100)If I connect to the very same pod via "kubectl exec -it" and run the command...
Celery can't find Redis in Kubernetes
As of Docker 3.x, the BuildError contains a new build_log variable which is a generator of output: try: return client.images.build(...) except BuildError as e: print("Hey something went wrong with image build!") for line in e.build_log: if 'stream' in line: logger.error(line['stream'].s...
If I build an image using the high-level docker-py sdk, I get a BuildError on failure., eg try: client.images.build(...) except: print("Hey something wrong with image build!") I know I can use the low level client API to directly hook in and stream logs, see How can I detect when docker-py client.build() fail...
How to get logs when build fails in docker-py?