Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
The default parallelism can be overridden during submission, or by the job itself. I would look in those places to see if the parallelism is being set explicitly.
I'm following this documentationhttps://nightlies.apache.org/flink/flink-docs-master/docs/deployment/resource-providers/standalone/kubernetes/about how to setup a standalone kubernetes session cluster using minikube. I've already installed minikube and I'm able to create pods. Inflink-configuration-configmap.yamlfile I...
parallelism.default not working for Flink Standalone Kubernetes Session Mode
Haven't tested, but my guess is that using this -https://github.com/schacon/simplegit-progitYou can do something like this - git log --pretty=format:"%s" > comments.txtAnd here's some info about the library -http://git-scm.com/book/en/Git-Basics-Viewing-the-Commit-History
I have a project on Git hub that I have been committing to frequently. I assumed that the comments I committed along with the files could be downloaded in some way. As a text file or even via the API but I can not see a way to download the comments.Is there a way I can download a text file containing all the commit com...
Github commit comments
Fllowing the documentation you can forward udp protocol with this line docker run -p 53160:53160/udp docker/getting-started So trying add this to your configuration file and should works. References: docker network
I'm trying to use the "Node.js & Mongo DB" devcontainer from Microsoft (https://github.com/microsoft/vscode-dev-containers/tree/main/containers/javascript-node-mongo/.devcontainer) with my udp server made in Nodejs. The problem is as follows: When i try to forward a port in devcontainer.json example : "forwardPorts": ...
How to forward a UDP port from a devcontainer?
AWS doesn't provide any metrics or logs or anything like what you are looking for with EFS. It's extremely difficult to deal with an issue like you describe. You will need to connect to each individual server that is using EFS and use the operating system's file system monitoring tools to determine how much each server...
I have an EFS that is mounted to File systems. Recently my burst credit started to decline until it got to 0 so whenever the efs is used the throughput utilization reaches its maximum (100%). This makes the EFS SUPER slow and almost unreachable. The solution for the issue was to change the throughput mode from burst to...
EFS is very slow - throughput utilization reached 100%
The YAML is valid. However, you are using a left double quotation mark “ like so: version: “2” Based on the error, seems like Docker Compose is not able to parse the version correctly. If you use a left double quotation mark instead of a quotation mark, the version which will be picked up by Docker compose will be “2”...
Here is the docker-compose.yml version: “2” services: web: build: . environment: MONGO_URI="mongodb://ravimongo:27017" ports: — “3000:3000” links: — ravimongo depends_on: — ravimongo ravimongo: image: mongo:3.2.6 ports: — “27017:27017” Here is the error: ERROR: Version in...
ERROR: Version in "./docker-compose.yml" is unsupported. You might be seeing this error because you're using the wrong Compose file version
I would recommend you to start fromscalingperspective. Separate apps that should be scaled from apps that shouldn't. That would bePods.Define the connection between apps taking into account fault tolerance and load balancing. Define the autodiscovery strategy. That's whatServicesis used for.Think about how you expose y...
I'd like to design my next project in a microservices-based architecture rather than as a monolith. Are microservices generally defined as a set of endpoints that compromise a service (i.e. common dependencies, functionality, etc.) or would each endpoint be its own service?I've usedServerless Frameworkbefore, which def...
Structuring backend (REST API) into microservices with Kubernetes (or serverless)
Adding your key to your ~/.ssh/config (using, as noted, a text editor) is only needed to: reference a private SSH key which is not a default one (as listed in "Checking for existing SSH keys"), like id_rsa, id_ecdsa, or id_ed25519. using a shorter SSH URL Meaning your config file can look like: Host gh Hostname gi...
Hello I am very new to using zsh terminal and am trying to set up my SSH key with github. According to this tutorial I have to open my /.ssh/config file and add my key. But after creating the file using touch ~/.ssh/config and then using the open command, the terminal goes opaque as if executing but then nothing happe...
zsh terminal can't open config?
It turns out all I needed to do was run Boot2Docker thentypeboot2docker ssh.This actually connects to the Unix VM, instead of running docker commands from a Windows shell.
I'm running Boot2Docker on Windows 7 (64 bit). Boot2Docker is supposed to be based on Tiny Core. I was trying to install docker-compose (a separate issue that I'm still working on), and several help topics (such asHow to install docker-compose on Windows) suggested using commands involving bothsudoandtce-load. Boot2...
Boot2Docker doesn't recognize "sudo", "tce-load"
If you'd like to set the IAM role that Packer uses during AMI creation from the command-line (e.g.from Jenkins), then you can use variables for doing so,e.g.using the following in your Packer script:"variables": { "packer_profile": "packer", ... }, "builders": [ { "type": "amazon-ebs", ... "iam_instan...
I'm currently building AMIs via Packer without a problem, but I am baking the AWS credentials into my scripts which is not what I want. Reading the Packer documentation they specify that if no credentials are found it can use an AWS role.I have created the policy and the role, but it's unclear to me how to tell Packer ...
How to use AWS roles with Packer to create AMIs
Your regex has anchor$more than once.You can use this rule:RedirectMatch 301 ^/listings/(cat/.+)/state/(.+)$ /directory/$1/region/$2ShareFollowansweredMar 25, 2016 at 11:43anubhavaanubhava771k6666 gold badges582582 silver badges649649 bronze badges0Add a comment|
I am trying to do a redirection from old structured links to the new version in htaccess, but I am having no luck.And the other answers in Stackoverflow dont help.So I have this link type of linkwww.example.com/listings/cat/shoes/state/nevada www.example.com/listings/cat/heels/state/arkansas www.example.com/listings/ca...
Change parameter names in redirect with htaccess
Make sure:your remote repo references the remote repo read-write addresshttps://github.com/DavidTunnell/delivery-driver-wage-calculator.git,check if GitHub did register an ssh key in your GitHub admin ssh section.That key should be namedGitHub for Mac - yourHostname.don't forget to add your GitHub username and email in...
I apologize if this is a newbie question. I have been working on trying to get a few projects on GitHub for hours.I am logged in to the application:But whenever I try to push or pull I get:I was initially trying to connect on Eclipse but was running into authentication errors as well. What am I doing wrong?
Github Authentication
GitHub's Pull Requests are designed to be per-branch, not per-commit. This is deliberate: After your pull request is sent, any new commits pushed to your branch will automatically be added to the pull request. This is especially useful if you need to make more changes. If you want to send ten Pull Requests in GitHub...
I forked a project. I worked on it for a bit. I now have ten different commits, each of which is independent of the other commits. I'd like to send each individual commit as a separate pull request, to let the maintainer of the upstream choose which ones he/she wants. I can't find an easy way to do this. The "easiest"...
How to send one pull request per separate commit?
Apparently I'm an idiot who can't read documentation and missed that part:In the same directory as yourpackage.jsonfile, create or edit an.npmrcfile to include a line specifying GitHub Packages URL and the account owner. Replace OWNER with the name of the user or organization account that owns the repository containing...
I have just published a private package on GitHub, trying to figure out how it should be working. now I'm trying to install it in another project. I authenticated withnpm login --registry=https://npm.pkg.github.comwith an access token that haswrite:packages,read:packagesandrepoprivileges. While trying to runnpm install...
Installing packages from github npm registry - auth error 401
You can use filters in the search bar to look for separate hosts.Use a query like > beat.hostname : abc and it will filter the log stream for just the hostname "abc"Tip : You can also add this hostname as a column in the log stream so that you can differentiate which log is coming from which host without even applying ...
I have installed ELK on my Ubuntu server and install filebeat on remote server-A and server-B. I have configured Log-stash to receive data from filebeat and forward them to Elasticsearch. Both servers logs are showing in Kibana-->obeverability-->Logs.The issue is both servers logs are got mixed and its hard to me find ...
How to show mulitple servers logs in kibana separatly from each other
Well, you could create a rule for taking the movie title and turning it into a slug. So, you'd know that you always lowercased the title, removed anything other than letters, numbers and dashes, and converted whitespace into a single dash.Then store that in another column in your database, and be sure you are forcing u...
I have had this problem for a while,Let say we have a movies websiteAnd we have a movie namedTest-movies123!in the database,now what I would do is make a URLwatch/test-movie123-{$id}/and then query DB with the ID,Now the issue with this is that the ID shouldn't be there, how can I go around this ? if I get thetest-movi...
Constructing URL without item ID and getting right item
@Scheduled(cron = "1 * * * * ?"): Runs your job at 01 second of every (any) minute.So you job runs once every minute.Test:@Scheduled(cron = "1 * * * * ?") public void run(){ System.out.println("Running Test Run : "+DateTime.now()); }Output:Running Test Run : 2016-06-08T20:06:01.019Z Running Test Run : 2...
I am using Spring 4.0.2 with Java configuration. There is not an XML configuration.I have@Scheduledannotation for a cron job. It is running twice. Can someone help me? This is what I'm trying.@Scheduled(cron = "1 * * * * ?"
@Scheduled is running twice with spring 4 java configuration
Sure, edit that constant in the source code and build own Grafana with own limit.https://github.com/grafana/grafana/blob/d4e4cb4c7179fd73a7f575d198dfed2ad9581a1a/pkg/tsdb/sqleng/sql_engine.go#L128Anyway, it looks like you are using wrong tool for CSV export. Grafana is not the best tool for MSSQL mass CSV export. I wou...
Hi I have a table view in Grafana which I am using for querying database of MSSQL which has over 1.5 million rows and trying to export it as a csv in grafana. I hit a roadblock where I couldn't retrieve rows beyond 1 million in grafana: "query row limit exceeded, limit 1000000". Is there any way out of this? I tried se...
How to get past Grafana open source million rows limit
I think what you have is probably the cleanest option. Another option, which I haven't tested, may be to set the VaryByCustom parameter and override GetVaryByCustomString in Global.asax. public override string GetVaryByCustomString(HttpContext context, string arg) { if (arg.ToLower() == “id”) { // Extr...
I have the following action: public class HomeController : Controller { public ActionResult Index(int? id) { /* ... */ } } I'd like to [OutputCache] that action, but I'd like that either: it doesn't use the cache if id == null; or it uses the cache if id == null but with a different duration. I think I can achi...
ASP.NET MVC OutputCacheAttribute: do not cache if a parameter is set?
2 This line COPY ./publish . will copy all publish files into your current work directory. The bin folder that contains your DLL now lives here: /app/bin/{your dll}. It is not necessary to then change the work dir to bin like you did; instead, you could either leave your w...
I am very new to docker. And I am trying to dockerize a helloworld API. But when I tried to docker run the image, I got this error: C:\Program Files\Docker\Docker\Resources\bin\docker.exe: Error response from daemon: container f891d7fc7af6e1183256043e6105fc87e25c6959d9745cc972b42c7b2e6f5a06 encountered an error durin...
Docker run gives "CreateProcess: failure in a Windows system call: The system cannot find the file specified. (0x2)"
Because this gets so much traffic, I'm adding the fully detailed code that you need to get this done. # Create a group and user so we are not running our container and application as root and thus user 0 which is a security issue. RUN addgroup --system --gid 1000 customgroup \ && adduser --system --uid 1000 --ingr...
I'm successfully running a simple dotnet core 2.1 web API application in docker but want to run it under a custom account instead of under root as this is supposedly best practice. I can add an account and change to that account, but then Kestral throws an error on startup. I've searched the web repeatedly and can't f...
How to run .NET Core 2 application in Docker on Linux as non-root
When you authorized with GitHub you probably did not explicitly grant access to the eventquarry organization.To fix this, first you need to revoke your token by logging into Github.com, then click settings > Authorized applications > Revoke "IBM Bluemix Toolchains". Now go back to Bluemix, and when you click on the Git...
I have a github account:https://github.com/remkohdevand am a member of several github organizationsI have added all organizations to the Bluemix DevOps Third Party applications authorizations.But I cannot change the default github organization to a different organization when I enable the Toolchain on Bluemix, so that ...
I cannot link toolchain to different github organization
sed is for doing s/old/new, that is all, for anything else just use awk. With GNU awk for the 3rd arg to match():awk ' match($0,/(.*msg=)([[:alnum:]]+)$/,a) { cmd = "xxd -r -p <<< " a[2] $0 = a[1] ((cmd | getline line) > 0 ? line : "ERROR") close(cmd) } { print } 'The above is assuming the syntax for callin...
I'm trying to use sed to properly parse the output of auditd records. These have encoded hex, long timestamps and UID/AUID which I need to decode/translate via commands.I am using pipes as I have to ship this across to the system journal. I have gotten this far:sed -r "s@msg=([A-Z0-9]*)\$@msg=$(xxd -r -p <<< \1)@"Sampl...
How to use sed to replace strings in piped input with command outputs using regular expressions
This update works for me well:Dockerfile-prometheus:... COPY ./prometheus.yml /etc/prometheus/prometheus.yml COPY ./prometheus.json /etc/prometheus/file_sd/prometheus.json ...docker-compose.yml:... prometheus: container_name: grafana_prometheus_ctnr build: context: . dockerfile: Dockerfile-prometheus volu...
I'm trying to configure prometheus and grafana to monitor my django app, but when execute docker-compose up command throws this error:grafana_prometheus_ctnr | level=error ts=2020-10-20T13:08:42.474Z caller=main.go:290 msg="Error loading config (--config.file=/etc/prometheus/prometheus.yml)" err="open /etc/prometheus/p...
Docker Prometheus Config
4 I found that by removing https:// from returned command: docker login -u AWS -p password -e none https://aws_account_id.dkr.ecr.us-east-1.amazonaws.com to: docker login -u AWS -p password -e none aws_account_id.dkr.ecr.us-east-1.amazonaws.com Then it works... This see...
Following instructions from AWS: $ aws ecr get-login returns command $ docker login -u AWS -p password -e none https://aws_account_id.dkr.ecr.us-east-1.amazonaws.com However, on Windows (this works on Mac) I receive error when trying to pull. $ docker pull aws_account_id.dkr.ecr.us-east-1.amazonaws.com/xxx Using def...
Pull an Image from Amazon ECR fails on Windows
So after some painful experience with AWS elastic beanstalk, here is my answer: (tl;dr we're switching to heroku) Beanstalk advantages: Autoscaling (but make sure you actually need this, it's easy to scale on heroku, it's just manual) Based on WAR format, if you're familiar with that already (but see below) Disadvan...
We're trying to decide whether to host our play! framework and mysql Java app on Amazon's Elastic Beanstalk or the new Heroku Java offering. I'm having trouble figuring out what the advantages of Heroku would be. One disadvantage is that Heroku doesn't auto scale nodes like beanstalk does. But are there advantages I s...
What are the advantages of the new Heroku Java support over Amazon Elastic Beanstalk
The simplest solution would be to actually use standard cron to execute yourwp-cron.phpfile at specific intervals. For instance, adding the following to your crontab would execute the Wordpress cron script every 5 minutes:*/5 * * * * /WORDPRESS_INSTALL_DIR/wp-cron.phpIf you setup a real cron job then I also think you s...
As you see in the title, I'm looking for an alternative to WP cron. As everybody knows (or should know): WP_cron only runs when people are using the wordpress site. F.e.: You setup a cron each 15 minutes, but nobody visits your site in the past hour, nothing happened.I'm looking for an other solution then server-cronjo...
Wordpress alternatives for WP cron
you need to also add exceptions forcdn.sstatic.netandajax.googleapis.com
I have problem when i open www.stackoverflow.com on my intranet which is protected by firewall, open very slow and only text appears, no image or format render on browser.my admin says that he has done setting in web blocker to allow this site, but if this site open any other url then i have to tell him then he will un...
need help to open stackoverflow in my intranet
In /etc/security/audit/events you have:* audit() AUD_It = printf "cmd: %d arg: %d"In /etc/security/audit/config add:... classes: others = AUD_It ... users: root = others user = others ...When you run the command:# audit start # audit query # audit shutdownYou will see:# tail -f stream.out event ...
Im hardening an AIX 7.1 box and i need to ensure that all starts/stops of "audit" system are logged on the AIX System.Does anyone have an idea on how can i achive this?
Log starts/stops events from Audit System
Use a category on the targeted class to override dealloc and set a breakpoint on it. http://developer.apple.com/iphone/library/documentation/Cocoa/Conceptual/ObjectiveC/Articles/ocCategories.html
I think i have an UIImage that has a higher retain count than it should have and i am probably leaking memory. I use this image as a thumbnail, to set a custom background to a uibutton. So the uibutton is holding a reference to it and so do i. But instead of 2, the retainCount is 3. Do i have to create a custom UIImag...
How can i see if dealloc is being called on a uikit object, or any object not created by myself
The ISP has Blocked the Port 22.Solved!
I am not able to ssh in git to github or bitbucket or gitlab.I typically have the following error message: how can I avoid it?========== output ============ssh -T[email protected]-i ~/.ssh/id_rsa -v OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016 debug1: Reading configuration data /etc/ssh/ssh_config debug1...
ssh connection timed out
git branch only shows local branches by default; use git branch -r to see remote branches or git branch -a to see all. git clone only creates one local branch, master by default, which tracks (again, by default) the master branch on the remote repository.
This question already has answers here: How do I clone all remote branches? (48 answers) Closed 7 years ago. If I do $ git clone https://github.com/dtu-compute/docker-mongodb $ cd ...
Why doesn't `git clone` clone all the branches? [duplicate]
In the Laradock Getting Started guide, it explains how to get Laradock running as a specified user: Note: You can add --user=laradock to have files created as your host’s user. Example: docker-compose exec --user=laradock workspace bash I believe this should solve your issue, as you will no longer have the Docker u...
I've newly started to use Laradock to build my Laravel projects but I have a problem in editing the files such as Controllers, Models, etc which are made by the php artisan command in the Laradock workspace. The reason is the user in the workspace is a root and on the other side, I'm trying to edit the file in my edit...
How to edit files which are made by Laradock workspace
Have you tried: git remote This will show the names of the remotes accessible from your repo. If you want to see what the names resolve to you can add the -v option for verbose output: git remote -v
I want to see all my remote repositories in Github with cmd. How can I do this?
How to see all remote repositories which were currently earlier configured in Github?
You can use a docker-compose.yml file to define and run multi container Docker applications. And then with a single command you can build and start all your services.You can run both Linux and Windows programs and executables in Docker. Docker creates thin virtual environments for your apps. Here is an example of what...
I found many tutorials about those but none uniting the 3 of them. I want to learn how to do that because I need to start deploying react websites I made and I wanted to deploy them on my domain and hosting on my computer for testing purposes. In what order should I learn those to archive that ? What tutorials do you ...
How to conect docker nginx with express and react app on windows
All looks fine here, seems that all you need is some patience. Its hard to say anything more without.nextflow.log.It might take hours depending on file size. Most likely process is running. Apart from log, you can open the second terminal and have a look at the CPU/mem usage (use e.g.toporhtop). You will see lot of CPU...
I'm on a Linux virtual-machine and I have installed Nextflow, created directories for the input files: index.ref and fastq and create a pipeline.nf script.Here below the code:#!bin/env nextflow params.index_dir="/mountpoint/nextflow_container/BWAIndex" params.ref="/mountpoint/nextflow_container/BWAIndex/human_g1k_v37_...
Mapping processing doesn't work using Nextflow scripting language
1 mysql -e "SET PASSWORD = 'YourChosenPassword';" Share Follow answered Jan 19, 2022 at 21:58 emadelbiehemadelbieh 34611 silver badge55 bronze badges 3 See "Explaining en...
I ran the command sudo mysql -e "SET Password=PASSWORD('rootpassword')" but it was showing the error .
Run sudo mysql -e "SET ***'rootpassword')" ERROR 1045 (28000): Access denied for user 'root'@'localhost' (using password: NO) on Github Actions
2 The nginx setting you are trying to use (/etc/nginx/conf.d/proxy.conf) is for Amazon Linux 1. Since you are probably using Amazon Linux 2 you should be using different files for setting nginx. For AL2, the nginx settings should be in .platform/nginx/conf.d/, not in .ebe...
I'm deploying with elasticbeanstalk with codePipeline. I have to increase file size be uploaded, so I add 02_nginx.config file is my app directory (.ebextensions). Here is my -2_nginx_config #Elastic Beanstalk configuration for 413 Request Entity Too Large:. container_commands: 01_reload_nginx:`enter code here` ...
I have a nginx client_max_body_size setting issue with elastic beanstalk
Solution: name in pip install command appears to be different:pip install awsiotsdkAnd nowimport awsiotworks.Is it normal for these two commands to use different names? I usedpipreqs .andpip install -r requirements.txtinitially, which makes the same mistake and assumespip install awsiotis what it needs to do.
I've looked up manyversionsof this problem, I don't think I'm falling for any obvious pitfalls, even using a virtual environment. Starting to wonder if there's something weird with this particular package.python3 -m venv venv source venv/bin/activate pip install awsiot ls venv/lib/python3.10/site-packages/ | grep awsi...
Python "No module named 'awsiot'" despite being installed
Because my var is array, with only 1 ID, (but it is still an array), therefore I should use[0]in the end of var.ShareFolloweditedMay 28, 2023 at 21:10Anonymous85322 gold badges55 silver badges2222 bronze badgesansweredMay 24, 2023 at 16:23Oleg ShalnovOleg Shalnov1Add a comment|
Mainly issue is that's i cant use apply_to in firewall with varsresource "hcloud_firewall" "appf" { name = "appf" apply_to { server = var.app-f-ids } rule { direction = "in" protocol = "tcp" port = "443" source_ips = [ "0.0.0.0/0", "::/0" ] } rule { direction ...
terraform hetzner firewall apply to node
Try to replace your rule with this:RewriteCond %{THE_REQUEST} ^[A-Z]{3,}\s/+index\.php\?route=common/home[?\s] [NC] RewriteRule ^ /? [R=301,L]
I am trying to forward /index.php?route=common/home to the domain name itself. I have tried the code below which works but it breaks the login page. I think the issue is that they used the same url for /admin/index.php?route=common/home and it will not allow me to login to the admin center. I have been unable to fig...
OpenCart SEO index.php .htaccess redirect
Map a PBOand decode/swscale directly into the buffer. After you unmap theglTexSubImage2D()becomes a GPU->GPU operation and nearly "free".With shaders you can render YUV directly and avoid using swscale for a YUV->RGB conversion, saving more CPU time for decode.
I am trying to decode and render super high res videos (4k x 4k) and the videos are encoded with h.264 codec. While decoding the video frames, we would like to render the frames on shaders, currently in opengl pipeline with glsl, so that we can manipulate the fragment colors on the fly. We have tested using FFmpeg deco...
Hardware Super High Res Video Decoding and Rendreing on OpenGL/DirectX
As of writingspatie/browsershotis the best solution for my problem. It supports Laravel 10 and no changes to code are needed to make it work. It does about whatnesk/puphpeteerdid and has an active community.
This packagenesk/puphpeteerhas been archived and doesn't work with Laravel 9 anymore. There are howeverforksthat I believe will work with Laravel 9 since the dependencies in the composer errors have been updated in the forks. I tried these steps in thecomposer docsbut can't seem to install a fork as a composer dependen...
PuPHPeteer php bridge package unable to install on Laravel 9
The nearest thing which comes to my mind is memoization for pure functions. Maybe also interesting for you might be this bookPattern Oriented Software Architecture Patterns Managementwhich has caching pattern in it.ShareFolloweditedApr 27, 2010 at 19:32BCS76.8k6868 gold badges190190 silver badges297297 bronze badgesans...
Given that it's one ofthe hard things in computer science, does anyone know of a way to set up a plugable caching strategy?What I'm thinking of would allow me to write a program with minimal thought as to what needs to be cached (e.i. use some sort of boiler-plate, low/no cost pattern that compiles away to nothing anyw...
Generalized plugable caching pattern?
Where it can be accessed hostconfig.json for a container can be found in WSL2 using Windows Explorer at the hidden network path \\wsl$\docker-desktop-data\version-pack-data\community\docker\containers\{containerid}\hostconfig.json Windows Explorer Screenshot with hostconfig.json Where it is not (but may be expected) T...
As I know, docker doesn't support any command that change port mapping after make container. But this answer said that by changing the hostconfig.json file, I can change port mapping. How do I assign a port mapping to an existing Docker container? I'm using docker desktop in windows + WSL2.So path of hostconfig.json i...
where is hostconfig.json (Docker desktop + WSL2 environment)
TableName : 'tablename', Key : { id: id}, ReturnValues : 'ALL_NEW', UpdateExpression : 'set someitem['+`index`+'].somevalue = :reply_content', ExpressionAttributeValues : { ':reply_content' : updateddata }array element edit via array index
I noticed that DynamoDB can add and remove items from an array but how do you search for an specific item inside an object if you want to update that one specifically? For example: In MongoDB you can search for someitem.$.subitem and update that specific item. Is there a way on how to do this with DynamoDB?Item: { s...
DynamoDB update inside an array of objects (nodejs)
you can extend ContainerAwareCommand instead and use$this->getContainer()->get('doctrine')->getManager()ShareFollowansweredNov 12, 2018 at 14:41Benjamin LemoineBenjamin Lemoine1001010 bronze badges3i think that works thx :) i can't accept the response for 9min so wait a bit a i put it to valid–APoorDevNov 12, 2018 at 1...
This question already has answers here:Symfony2 - How to access the service in a custom console command?(2 answers)Closed5 years ago.Hey guys i want to know how can i do a cron tab but that cron tab has to update database so i need$this->getDoctrine()->getManager()But for this i need to extendControllerbut i already ex...
Symfony - call Doctrine in cron tab [duplicate]
I created a mini cluster usingkubeadmwith two ubuntu18.04 VMs in one public network.Indeed changing IP address of the worker node doesn't affect the cluster at all unless new IP address doesn't interfere with--pod-network-cidr.Kubeletis responsible for it and it uses several options:The kubelet is the primary "node age...
I have tried to change the NIC ip of the worker node directly. It seems that the master node automatically updates the ip information of the worker node. And it does not have any negative impact on the kubernetes cluster. Is it the simple and correct way to change the worker node ip? Or are there some other important s...
How to change the kubernetes worker node ip address correctly?
I have a suspicion that it has something to do with user - not much help from the error message.@TietjeDK is correct that it is just a misleading error message. It means one of two things is happening (or maybe both): thekubectlbinary is newer than the supported version range of the cluster (so: using a v1.11 binary ag...
I'm trying to update an image in Kubernetes by using the following command:kubectl set image deployment/ms-userservice ms-userservice=$DOCKER_REGISTRY_NAME/$BITBUCKET_REPO_SLUG:$BITBUCKET_COMMIT --insecure-skip-tls-verifyBut when I receive the following error:error: the server doesn't have a resource type "deployment"I...
Kubernetes set image missing resource type 'deployment'
If you are calling from your application to the gmail server. That would be an outbound rule so you need to add a outbound exception to the firewall.ShareFollowansweredMar 18, 2022 at 10:39radulfrradulfr11433 silver badges1313 bronze badges2If you don't care what the server is and only need it for dev, I highly recomme...
hi I'm trying send email , but I just get this error :an attempt was made to access a socket in a way forbidden.I'm usingsmtp.gmail.comI search about this and I find it's bc firewall don't allow to use port 587I tried to addInbound rulesin firewall , but did not answer.so anyone know how can enable port 587 to send ema...
problem with port 587 how enable this for send email
You can use this in your.htaccess:RewriteEngine On RewriteCond %{HTTP_HOST} ^subdir\.example\.com$ RewriteCond %{REQUEST_URI} !^/subdir/ RewriteRule (.*) /subdir/$1Make sure you clear your cache before testing it.A suggested edit from OP:RewriteEngine On RewriteCond %{HTTP_HOST} ^(.*)\.example\.com$ RewriteCond %{REQ...
This question already has an answer here:Difference between $1 vs %1 in .htaccess(1 answer)Closed7 years ago.All my rewrite rules up to this moment used $1, $2... variables to form a clean URL.Now I need to redirect basing on a part of domain name (http://subdir.example.com=>http://example.com/subdir) and many examples...
.htaccess Rewrite rules: what is the differences between %1 and $1 variables (percentage vs dollar sign)? [duplicate]
4 To make it simple, anytime you make a new, you should make a corresponding delete. In your case, you have to delete the allocated objects Obj at latest before your objects Objects are deleted, which means in the destructor of Objects, which also means that you have to kee...
I want to know when or if I have to delete this object. Here is the constructor of a basic class Object Object.cpp: Objects::Objects{ Obj one = new Obj; Obj two = new Obj; } I know when allocating memory you are supposed to delete it at some point, but I have allocated memory in the constructor and want to us...
Allocating memory - Pointers
Those "large exponential values" you are seeing are cumulative. I.e. the CPU (or all CPU cores?) have been idle for 1000+ hours (4.25e6 / 3600) since the VM was started, so they look very reasonable.The reason you are getting negative values is because of sampling. In theory all samples are exactlyscrape_intervalsecond...
I have set up a monitoring system using prometheus, with AWS EC2 Auto discovery, and node exporter. Using the following formula to get CPU Utilization:100 - (avg by (instance) (irate(node_cpu_seconds_total{instance="instancexyz" ,mode="idle"}[5m])) * 100)However, in one particular ASG, I am getting CPU percentage in la...
Node Exporter shows Wrong Idle CPU time
Here is a working solution. You simply have to define a variable for the container outside the main pipeline. Then you can use it anywhere in the pipeline to start or stop the container. In particular, you can remove the container in post{ always{ } }. def receiver_container pipeline { agent any stages { ...
I have a Jenkins pipeline that builds and runs a Docker machine, not as an agent, but using a scripting block along with the Docker Pipeline Plugin methods docker.build() and Image.run(). This works fine but if the build fails, the docker container is left running! I currently have Container.stop() in a post{ always{}...
What is the proper way to stop and remove a docker container in Jenkins when using Docker Pipeline Plugin?
So I know nothing of grafana but I took your description and query to come up with this which runs on my historian. Mod to fit your needs I guess?SET NOCOUNT ON DECLARE @StartDate DateTime DECLARE @EndDate DateTime SET @StartDate = DateAdd(mi,-5,GetDate()) SET @EndDate = GetDate() SET NOCOUNT OFF SELECT QueryDate, ...
I pulled 3 time series signals A, B, and C from a SQL server data source. I want to implement the following logic:if A > 0: if B > C: Status = 1 else: Status = 0 else Status = 0What is the best way to implement this logic in Grafana? I tried Transform in Grafana and it is possible there to subt...
How to implement this logic in grafana dashboard?
honestly i would suggest going thru the github tutorial, its quick and simple. answers most of the basic questions you'll have. after you have your files setup in your working directory, open github, and either create a new branch from your master. at that point will be able to select the new branch and commit your ...
I´ve just started with Git and build my first project with it. I´ve installed GitHub Dektop too and can sync this with GitHub.com. I read a lot in the extensive help of GitHub. Now i know, how to put my files into GitHub. But i don´t find anything about how to get files from the repo, i.e. to send a link to others. Ho...
how can i download my files from a git repository
3 I am unsure if the icons Github utilizes are under creative commons. However, the icon packages below are used in other popular software packages that might be a fit for your project (e.g. you can always convert to grayscale and resize them using Gimp): famfamfam icons fu...
I'm talking about the icon set that Github uses for their own site design. Example Page On this page, near the top right corner, the "Watch" and "Fork" buttons, as well as the buttons next to them have beautiful icons! Any idea where the full set can be found? To clarify, I'm just looking for minimalist grey icons of ...
Where can I find the icon set Github uses?
2 In our org, we ended up going with an internal-only ALB and we enabled MultiValueQueryStringParameters to pass data into the Lambda function and to execute it. This is the only way I could find to provide an internal-only URL that I could further protect with a securit...
I have a lambda function in AWS inside a VPC. I want to attach http handler (function URL). The problem is, if I enable the function URL then it creates a public endpoint. Alternatives I don't want to use enable AWS_IAM security (then the caller will need to use AWS SKD and get token and all) API gateway trigger (I a...
AWS Lambda function via Function URL invoke only within VPC
If you setservice.spec.externalTrafficPolicyto the valueLocal, kube-proxy only proxies proxy requests to local endpoints, and does not forward traffic to other nodes.kubectl patch svc servicename -p '{"spec":{"externalTrafficPolicy":"Local"}}'If there are no local endpoints, packets sent to the node are dropped.Forclus...
I have a Kubernetes cluster of 3 nodes.A sample deploymentapiVersion: apps/v1 kind: Deployment metadata: name: nginx-deployment labels: app: nginx spec: replicas: 3 selector: matchLabels: app: nginx template: metadata: labels: app: nginx spec: containers: - name...
Disable Kubernetes replica set load balancing
1 First, I think you should plug your thumb drive into another computer to confirm that your thumb drive is available. And if it is normal, there is something wrong with the Windows 10 system image tool. I recommend a perfect backup software - AOMEI Backupper which can crea...
Fellow Forum Members, How does one create a System Image of an entire Windows10 installation onto a 128GB thumb drive? I tried doing it through the Windows 10 System Image tool but it sees a thumb drive as an invalid storage device. Then I Googled the subject and learned I need to convert the thumb drive to a Local D...
Windows 10 Image Backup to Thumb Drive
2 A tag-name is meant to be a specific name for one specific commit, i.e., one raw hash ID. When you modify the README.md file (or add a new one) and make a new commit, that's ... a new commit. It has a different hash ID. You should make a new tag, if you want to use a ta...
In case I want to edit my README.md of a tag, however when I give the following commands nothing happens: git checkout <tagname> git add README.md git commit -m "MESSAGE" git push origin <tagname>
How to change files from a specific TAG already in the repository?
DNS-SD and mDNS are available in a new extension to Windows.Devices.Enumeration in Windows 10.https://msdn.microsoft.com/en-us/library/windows.networking.servicediscovery.dnssd.aspx(Side note to these slides: The watcher.Updated event handler should be set in order for watcher.Added to trigger.)
It used to be that Windows would only resolve multicast DNS addresses (say,example.local) if iTunes was installed. Is that still true or has that changed in recent Windows versions? (I gave up on Windows years ago and so I cannot check easily myself.)
What is the current state of mDNS on Windows?
What causes thisI believe this is a bug.In some cases, this can occur if you try toterminateorrebuildyour Elastic Beanstalk environment and the operation fails to delete yourAWSEBSecurityGroup.There are reports (see comments) of other causes besides this.How to fix itThe AWS documentHow do I terminate or rebuild my AWS...
I'm currently using the "Ruby 2.6 running on 64bit Amazon Linux 2/3.0.2" image, and by looking, inside the EC2 instance at the /var/logs/eb-engine.log ("eb logs" command won't show me this), there is a recurring error:[ERROR] failed to parse JSON file /opt/elasticbeanstalk/deployment/app_version_manifest.json with er...
AWS Elastic Beanstalk won't deploy my Rails app even once
Is there a reason why you are usingMappingin between?You could easily use!SubinsteadResources: EC2Instance: Type: AWS::EC2::Instance Properties: InstanceType: !Ref InstanceType KeyName: !Ref KeyName Tags: - Key: Name Value: Test UserData: Fn::Base64: ...
I have this under parameter section ,Parameters: PlatformSelect: Description: Cockpit platform Select. Type: String Default: qa-1 AllowedValues: [qa-1, qa-2, staging, production]I need to reference this value in my UserData. I’m using Mappings in between.Mappings: bootstrap: ubuntu: print...
Reference Parameter Value in UserData in AWS Cloudformation
I ran into a problem with the script above. Unfortunately let's encrypt doens't override fullchain and privkey but adds new versions when certificate is due to renew: fullchain2.pem privkey2.pem So I had to alter the script accordingly. I also put the renew and nginx part inside so we need only one cronjob: #!/bin/bas...
I've got a parse-server up and running on digital ocean following this guide. When configuring mongo db for migration you execute this command: sudo cat /etc/letsencrypt/archive/domain_name/{fullchain1.pem,privkey1.pem} | sudo tee /etc/ssl/mongo.pem After that the tutorial says: You will have to repeat the above com...
cron jobs for let's encrypt ssl renewal with mongod on nginx
Using a reverse proxy will be the solution, as described here:https://learn.microsoft.com/en-us/aspnet/core/host-and-deploy/linux-nginx?view=aspnetcore-6.0Quote from the article:Kestrel is great for serving dynamic content from ASP.NET Core. However, the web serving capabilities aren't as feature rich as servers such a...
I run a .Net 6 API on my Raspberry PI. I can get data from this API using the browser “https://192.168.178.51:7001/swagger/” . Also getting data from entering the url “https://192.168.178.51:7001/api/status/” returns the expected values. So the API is working so far. But when I try to get data from my Blazor WASM App u...
.Net 6 API - can't get data -> ERR_CERT_AUTHORITY_INVALID
As far as I know, the intentionisto replacegit checkoutwithgit restoreandgit switchas appropriate.Wheregit checkouttakes two positional arguments to specify the branch and a file within that branch,git restoreuses the--sourceargument to specify the "tree" (could be a branch, tag, commit, etc) from which to take the fil...
I am trying to rebuild my git workflow by usingrestoreinstead of the oldcheckout. I knew that restore canundomy previous work (on my branch). But what if I want to checkout a file from another branch? Thecheckoutversion is quite simple,git checkout main -- my-fileCan I do the same thing withrestorecommand?My further qu...
How to git checkout a file on another branch using "restore" instead?
According to theGitHub API documentationfor creating a release, below is an example requestBody{ "tag_name": "v1.0.0", "target_commitish": "master", "name": "v1.0.0", "body": "Description of the release", "draft": false, "prerelease": false }But in your code, you have formed the requestBody in the below for...
I am trying to publish the github release notes via Java REST service. The have achieved the same by writing the sh script and using curl post call.I have written the code to do POST call using HttpUrlConnection by passing the JsonObject data.String postUrl = "host_name/api/v3/repos/"+ userName + "/" + project_name ...
Not able to create github release notes using http post call
A given CUDA version (e.g. 5.0) has a minimum GPU driver version associated with it. Any newer version should work, but versions older than the minimum driver version will give an error, as you are seeing.based on my memory:cuda 4.0: 270.41.19cuda 5.0: 304.54cuda 5.5: 319.37The above may be approximate, but it gives y...
My run time system gives an error "CUDA driver version is insufficient for CUDA runtime version" when I try to run CUDA programs on Mac OS X 10.6.8 installed with CUDA ToolKit 5.0 (CUDA driver: 5.0.45) and GPU driver: 1.6.36.10 for GeForce 900M. I was unable to locate the correct (updated) GPU drivers for GeForce 900M ...
CUDA on Mac OS X 10.6.8, GeForce 9400M
Docker as technology can only be run on Linux OS. Docker Desktop does 2 things. creates Linux VM on your host OS (Windows / Mac) forwards docker-cli commands from the host OS to created Linux VM --> due to this the containers are created on Linux VM, and on the surface it appears to be working on Windows/Mac Thus we...
As a technology, we understand the docker creates virtualization and there is a docker hub, where Docker images are stored.To explain in simple words, what purpose does docker desktop serve?Or why do I need to install Docker Desktop App to use DockerIf we need to use docker on linux server, do we need to install Docke...
What is the use of Docker Desktop?
If you only want to run the job once, don't usecron; it has no way to achieve that.If you want to run it at a later time, possibly after you've logged out, useat(1).If you want to run it after a short delay, like 1 minute in your example, and you plan to stay logged in (or the job makes itself a daemon), it may be simp...
I need to write a cron job to write a Python script, let's call it "script.py" and the command would bepython3 script.pyPlease note: It's a script that continuously runs and if I set it to rerun, it would spit out the script every time it would be scheduled like if I put it for every minute. I tried to schedule the cro...
How to run cron job now and never run it again or run it every month?
Try adding--before the--runlike this:cmd /C C:\php\php.exe "C:\cron\cronrun.php" -- --run >> "C:\cron\log\cronrun.log" 2>&1The clue is in the usage message:$ php --help Usage: php [options] [-f] <file> [--] [args...] ... args... Arguments passed to script. Use -- args when first argument ...
I want to run a PHP file via Task Scheduler on Windows. Everything runs except a parameter comes in.This runs:cmd /C C:\php\php.exe "C:\cron\cronrun.php" >> "C:\cron\log\cronrun.log" 2>&1This is not running:cmd /C C:\php\php.exe "C:\cron\cronrun.php" --run >> "C:\cron\log\cronrun.log" 2>&1How can I run a job wh...
PHP cronjob on Windows with parameter and pipe
1a) Nginx is asynchronous server (event based), with single worker itself they can handle lots of simultaneous connection (max_clients = worker_processes * worker_connections/4 ref) and still perform well. I myself tested around 20K simultaneous connection on c1.medium kind of box (not in aws). Here you set workers to...
I'm building an asynchronous RESTful web service and I'm trying to figure out what the most scalable and high performing solution is. Originally, I planned to use the FriendFeed configuration, using one machine running nginx to host static content, act as a load balancer, and act as a reverse proxy to four machines ru...
What's the most scalable and high performing Amazon Web Service (AWS) configuration for a RESTful web service?
5 RunInstances:SV001 means you are running EC2 spot instances in VPC zone #1. Relatively, SV002 could mean running EC2 spot instances in VPC zone #2. By far there is no documentation stating the meaning of those values, and AWS may change that at any time. Just for your ...
My AWS billing report shows for example RunInstances:SV001 under lineItem/Operation. Meaning of RunInstances is obvious but what is the SV00? about? Thanks
What does RunInstances:SV00? under lineItem/Operation mean in AWS billing report?
A typical pattern here is to use alog processor sidecar. Some of the ones you can use are:Fluent-bitFluentdLogstashlogspoutFilebeatFor any logger about you can use aConfigMapfor different types of applications.Another pattern is to use any of the above tools deployed asDaemonSetwhere each daemon in the Kubernetes picks...
I am trying to find out if I can use a container environment variable inside the Log4Net.config to separate our logs environment, for easier reading on Kibana.My idea was to set a variable in the container that appends into the ApplicationName value of our RabbitMqAppender.I probably could use a configMap for the Log4N...
Is there a way to use Kubernetes Container environment variables in the Log4Net.config?
Open the main GitHub projecthttps://github.com/pytorch/pytorchThen type in this page the keyboard shortcutt: you will enter the file finder mode (introduced in 2011). (https://github.com/pytorch/pytorch/find/master)Type "chunk.h": the first result will be your file.If the file is not the right one (header but not sourc...
I can’t seem to find the source code for torch.chunk in PyTorch’s Github page or in the documentation.Anyone knows where this is in PyTorch’s Github page?
PyTorch: `torch.chunk` source code Github location
According to dive into HTML5, the page that references the manifest is automatically included in the manifest.http://diveintohtml5.ep.io/offline.htmlQ: Do I need to list my HTML pages in my cache manifest?A: Yes and no. If your entire web application is contained in a single page, just make sure that page points to the...
The main page of my mobile web app is a .jsp page. My app requires login (Google App Engine), so there is a Log In button when the user is not logged in and a Log Out button when the user is logged in, all handled by code on the .jsp page.I load a lot of JS code on the page, so I used a .appcache file to cache that. Un...
webkit .appcache file caches dynamic page
Git doesn't store your credentials as part of the repository (with one exception). How it gets your credentials depends on how you clone it. If you do it using HTTPS with your username and password hard coded, it will always use that username and password. Don't do that. It will write your password in .git/config. Tha...
I'd like to clone a Github repo using a generic Github account, and then allow another user to use their personal account to push to the repo. Is there a way to do this? Will the initial clone associate the generic account with the local repo, and if so, is there any way to remove that association and for the user to ...
How can I clone a git repo with one account and then push with a second one
best way is first time you login into server save your --servercert in clipboard like thisecho "password" | sudo openconnect -u username uk2.cisadd.com --servercert sha25:xxxxxxx
Is there a way to store the server key after a successful connection, the way ssh does?No matter how many times I connect, I have to always type "yes" to accept the server key. I wish it would be accepted and stored forever.############################### amir@amirpc:~$ sudo openconnect uk.cisadd.com -u myusername PO...
how to store server key permanently? openconnect
I think if you are usingGPUArraysthe way to copy from device to host is with the.get()method. For exampledist = np.zeros(num_points).astype(np.float32) dist_list = [gpuarray.to_gpu(dist) for i in range(100)] buf = dist_list[0].get()
I'm copying via pycuda some arrays on the GPU and then store the pointers to these arrays. How do I recuperate the data back?dist = np.zeros(numPoints).astype(np.float32) distAddress = [gpuarray.to_gpu(dist).ptr for i in range(100)]If I call the memcpy_dtoh function:buf = np.zeros(400).astype(np.float32) cuda.memcpy_dt...
Copying arrays with pyCuda from the GPU using pointers
You installed theMemcachedclient (not theMemcacheclient, which is, very confusingly, also aMemcachedclient). Use theMemcached classinstead.In case you're wondering what the difference between the two clients is: here's a nicecomparison table.
Just finished installing libevent(1.4.8), memcached(1.4.5), pear, and libmemcached(0.40) to my lamp server (running PHP 5.2.10 & Centos 5.5 Final), and as far as I can tell, everything installed correctly (was able to address all errors during installation).However, after finally getting everything updated and installe...
Memcached installed (In theory), PHP unable to use memcache_connect()
A maintained alternative ishttps://github.com/adnanh/webhookwhich allows you to do install local webhooks with scripts attached.Example config:- id: redeploy-webhook execute-command: "/var/scripts/redeploy.sh" command-working-directory: "/var/webhook"Default port of the webhook process is 9000, so the following URL...
Please let me know if I can execute a shell script on the same server as Prometheus/alertmanager on an alert trigger? If so, help me with the configurations.
Execute shell script through prometheus alert manger
Have you looked at this articleDockerizing an Apt-Cacher-ng Servicehttp://docs.docker.com/examples/apt-cacher-ng/extractThis container makes the second download of any package almost instant.At least one node will be very fast, and I think it should possible to tell the second node to use the cache of the first node.Ed...
I have two hosts and docker is installed in each.As we know, each docker stores the images in local/var/lib/dockerdirectory.So If I want to use some image, such asubuntu, I must execute thedocker pullto download from internet in each host.I think it's slow.Can I store the images in a shared disk array? Then have some h...
How to share images between multiple docker hosts?
Based on your comments, I'd suggest that you split your task into two parts:1) In part 1, parse the JSON files using regexes and generate two CSV files in simple format: no headers, no spaces, just numbers. This should be quick and performant, with no memory issues: read text in, write text out. Don't try to keep anyth...
I'm working on an app that processes a lot of data..... and keeps running my computer out of memory. :(Python has a huge amount of memory overhead on variables (as per sys.getsizeof()). A basic tuple with one integer in it takes up 56 bytes, for example. An empty list, 64 bytes. Serious overhead.Numpy arrays are great...
Memory-efficient 2d growable array in python?
The latest AWS Application Load Balancer (ALB) can do the trick. This works for meFollow the steps here,1. Set up the ALBAWS documentation here, Follow steps up until Listeners tab,https://docs.aws.amazon.com/elasticloadbalancing/latest/application/tutorial-load-balancer-routing.html2. On the Listeners tab,2.1. Modify ...
I have a working domain, x.y, which is tied to an EC2 instance on a VPC.I want a path on that domain, x.y/z, to be routed to an external, non-AWS (IPv4*) microservice. Can this be done with ALB?I have followedUse Path-Based Routing with Your Application Load Balancerto set up target groups, but can't seem to link them ...
Path-based routing to external resource with AWS Application Load-Balancer (ALB)
) Yes, it is configurable within firehoseKinesis Firehose buffers incoming data before delivering it to Amazon S3. You can choose a buffer size (1-128 MBs) or buffer interval (60-900 seconds); whichever condition is satisfied first triggers data delivery to Amazon S3.Creating an Amazon Kinesis Firehose Delivery S...
When writing records to an AWS Firehose which is configured with S3 as the output destination, how long is this data buffered before it is written to S3? Or is there a minimum size threshold?For example, I'm doing the following to add records:aws cli:aws firehose put-record --delivery-stream-name mytestfirehoseafds --r...
AWS Firehose buffering
if you want to use just native ooptions of kubectl commandkubectl get nodes <nodename> -o jsonpath='{.status.capacity.pods}{"\n"}'If you dont need trailing newline character after the output:kubectl get nodes <nodename> -o jsonpath='{.status.capacity.pods}'
I am trying to get the maximum capacity of pod per node. I am runningkubectl describe node nodenameand trying to grep the pods limit in capacity section. Any help would be appreciated. The output is like this.Capacity: attachable-volumes-azure-disk: 8 cpu: 4 ephemeral-storage: ...
Get specific attribute from kubectl describe node
To get the name of the repo in the curernt working directory, rungh repo view --json name -q ".name"This extracts the name of the repo from the command that lists information about the repo in the current directory, and extracts the actual name from the JSON output. For instance, if you were in the working directory fo...
Use GitHub cli (gh), how do I get a current repo name?I can get a listing of the repos I have but how do I get the current repo?
GH cli - how to get current repo?
The simplest option is to clone the target repo, copy the files into the target repo, use the git commandline to stage the files and then commit them. Add the code below in a script step run: | git clone https://.:${{ secrets.GITHUB_TOKEN }}@github.com/project target rm everything but the .git directory co...
my company has created a complex way to move files between environments and now we would like to move certain built JS files (transpiled and minified) from one github repo to another. Is this possible using github actions?
github action to copy a file from one repo to another
This should work:RewriteCond %{HTTP_HOST} ^yourdomain\.com$ [OR] RewriteCond %{HTTP_HOST} ^www\.yourdomain\.com$ RewriteRule ^images\/?(.*)$ "http\:\/\/cdn\.yourdomain\.com\/$1" [R=301,L]However, please note that this is just a temporary solution!In order to get the max out of your CDN, you need manually point images t...
Recently I bought a CDN and set it up. In my site, the images are stored in a folder named 'images' and the Image urls are obviously linked in this manner. (*Ex : images/some_image.png*)Since I want to use the CDN right way, I need to rewrite the urls without having to manually change each and every image path.I tried ...
CDN Related : Rewrite Image URLs automatically from .htaccess
bundler is not actually hanging, it is downloading the content of the repository in the background.The repo-content seems to be of a significant size (multiple megabyte). Depending on your connection you will have to wait a while for bundler to finish downloading. Afterwards it should install the gem as usual.Unfortuna...
I have found fork of one gem in github, which is more up to date. I wanted to installthis forkto my rails project. So I have added this line of code to myGemfile:gem 'rootapp-wkhtmltopdf-binary', github: 'Root-App/wkhtmltopdf_binary_gem'When I runbundle install, it just hangs at fetching.Why does this happen?
Rails, trying to install forked gem from github hangs bundler
If you pull from the home computer, it can do one of the following things - Successfully pull changes from remote to local if git finds out there are no conflicting changes. Give an error if changes from remote can not be taken directly since there are conflicting changes in your local branch. In the later case, you...
For some reasons I'm working on a same project in two different computers. I would like to work in one of them, save the changes to a commit and push it to a GitHub repository. When I get home, if I want to keep on working on the project I just want to pull it from GitHub and work from where I left it. The problem is ...
GitHub - Pull localy an existing project and replace existing files without removing non existing one
It seems, that you created keystore in BCFS format using the Bouncy Castle library. If you want to see it in human-readable format, you can specify-storetype,-providerand-providerpathoptions for keytool:keytool -keystore keystore.bckfs -storetype BCFKS -providerpath "bc-fips-1.0.2.jar" -provider org.bouncycastle.jcajce...
I have a jks file and when i run the commandkeytool -keystore db-ssl-truststore.jks -list, i get the errorkeytool error: java.io.IOException: BCFKS KeyStore corrupted: MAC calculation failed.
Getting "BCFKS KeyStore corrupted: MAC calculation failed." error
-1To display the current branch you're on, without the other branches listed, you can do the following:git rev-parse --abbrev-ref HEADReference:hereor with Git 2.22 and above:git branch --show-current
Does anyone know how to view the branch name? Even on manually run workflows? Can't see any obvious settings to change it.*** EDIT ***Here's how it looks in most repositories (with the branch name in blue):And here's the exception repo with no branch name link:
Github doesn't displays the branch name next to the workflow run of manually run workflows
Then leave a comment for the request asking them to re-submit the pull request using the new branch and that I'll be closing the current requestThey don't have to re-submit, it you test and merge first locally, as described in the "Merging a pull request" GitHub page.git checkout master git pull https://github.com/othe...
Say I have a repo and someone forks it. Then they do work and submit a pull request. But the code contains a large number of lines and/or it creates a GUI. I'd like to fetch it so I can actually see it run from Eclipse before merging it into my master. The options I've come up with are:Create a second repo in EGit and ...
Github / EGit pull request workflow
You need to place this line on top of your .htaccess to turn offMultiViewsoption:Options -MultiViewsOptionMultiViewsis used byApache's content negotiation modulethat runsbeforemod_rewriteand makes Apache server match extensions of files. So/filecan be in URL but it will serve/file.php.
I have a strange problem. Here are two examples of a URL rewrite in my htaccess file. The first one doesn't work and returns the variable for "query" as null in the PHP file.RewriteRule ^search/([A-Za-z0-9-\s]+)$ search.php?query=$1The second one (that I have used as a test) works perfectly fine.RewriteRule ^foo/([A-Za...
htaccess rewrite doesn't work when path has same name as file
A Service of type LoadBalancer is a declaration of your desire, and it should cause a DigitalOcean LB to be provisioned. Implementing this desire is the responsibility of a cloud controller: code which subscribes to the list of Services of type LoadBalancer and uses the cloud provider's API to ensure a cloud provider L...
I have created K8S cluster withRancher on DO(DigitalOcean)platform and behind the scene, it creates VPS(Droplet) to build the K8S cluster.The way, how I have created on Rancher:and choosed the DO pluginOn DO platform, you have the opportunity the create Load Balancers to add to K8S cluster, as described onhttps://www.d...
How to assign an Load Balancer-IP to NGINX ingress?
Try to go to Settings -> General -> WordPress Address (URL) and remove the trailing slash.
I can't remove the trailing slash from homepage linkexample.com/blog/. I tried it from.htaccessas well as from site config.RewriteRule (.+)/$ $1 [R=301,L]It is working from permalinks, and all categories and posts being without trailing slash. But homepage still has it and is really annoying and not good for my SEO.** ...
Wordpress - trailing slash removal
Redirect by PHPheader("Location: server.myserverdomain.com/~seconsite/blog"); die();Redirect by HTMLMethod 1 :<meta http-equiv="refresh" content="0; url=http://server.myserverdomain.com/~seconsite/blog/" />Method 2 :<script>window.top.location = 'http://server.myserverdomain.com/~seconsite/blog';</script>
I have a wordpress installation that is located in the root folder: mydomain.comThis website has a blog that is directed to mydomain.com/blogWhat I would like to achieve is to redirect all traffic that goes to mydomain.com/blog to another wordpress installation that has a blog and located on a temporary domain on my s...
Redirecting a wordpress blog
Expires modification sets the cache to expire a certain time after the last modification of the file. Even if your file is modified again within the set cache time-limit, clients that have already cached the file will not look to find the new version.Apache Expirationcontains information on the expire module.Here's an...
I'm using an .htaccess file to set the headers of a website. I want to know how to set the files cache to expire when after a modification has been made on them. Is this possible?i.e. Will this work?ExpiresDefault "modification"In other words, I want the cache to stay until the files have been modified from the server.
How do you set the cache to expire after modification?
Before investing much time and effort, you may want to take into consideration that Amazon is promising a High-Availability offering on top of the Amazon Relational Database Service (RDS). New Features for Amazon RDS Coming Soon High Availability Offering — For developers and business who want additional resilien...
I'm currently spec'ing out a solution running on EC2. Setting up web servers, utilizing S3, Cloud Front, Cloud Watch, etc have been straightforward enough. Using Elastic Load Balancers for HA cross Availability Zone for our web servers has also been straightforward. I would really like to have cross Region Availabilit...
Amazon EC2 High Availability Database Architecture