Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
As you can see in docker settings, docker -> settings -> resources -> Disk image location , images are stored inside your home directory. Removing this directory resolves this issue. rm -rf ~/Library/Containers/com.docker.docker
I have renamed Mac Os user and home directory by https://support.apple.com/en-us/HT201548 After this Docker Desctop doesn't work(error). I have tried to reinstall Docker Desktop several times, but the error is the same. Docker Desktop is trying to create a directory under my old user directory which doesn't exist. I d...
Docker does not start if the home directory is renamed on Mac
There are four major mechanisms available for you (Lazy comes in 4.0, so it is no option) lazy initialization virtual proxy ghost value holder each has it own advantages. i suggest a value holder, which populates the dictionary on the first call of the GetValue method of the holder. then you can use that value as ...
Assume that I have the following object public class MyClass { public ReadOnlyDictionary<T, V> Dict { get { return createDictionary(); } } } Assume that ReadOnlyDictionary is a read-only wrapper around Dictionary<T, V>. The createDictionary method takes significant time...
Memory management / caching for costly objects in C#
It turns out thatnpm installmight take too long to run, cuz once I put node_modules into the zip and removenpm installfrom Dockerfile, it takes 3-5 minutes for deploying now.ShareFollowansweredSep 25, 2016 at 7:39Jerry ShenJerry Shen3311 silver badge33 bronze badges3Is it really the couse? Or you had some problem with ...
Has anyone encountered failed deployment when deploying docker app to aws eb?Here's a piece of logtime="2016-09-20T09:36:42.802106539Z" level=error msg="Handler for DELETE /v1.23/containers/c7bc72d9ccec returned error: You cannot remove a running container c7bc72d9ccec6557ddca8e90c7c77b350cb0c80be9a90921478adccd70a2b97...
AWS Elastic Beanstalk Docker deployment failed
Assuming your tree looks like this: A---B---C---D master If you checked out commit A, then you are currently in detached HEAD state, meaning there is no current branch. You cherry-picked commit D, but you're still in detached HEAD state: A---B---C---D master \ D' (no branch) Here D' is...
Recently I read about cherry picking from this I got that, cherry picking in git means to choose a commit from one branch and apply it onto another. Now I have to scenario as say I have four commits as A, B ,C and D. Currently I am on commit A and by using git cherry-pick <hash_of D> I pick-up changes of commit D. But...
create a separate branch along with git cherry-pickup
There are some webhooks to be added to your repository server, seehttps://packagist.org/about#how-to-update-packagesfor them - this will update the list of package versions immediately on pushing or tagging. Otherwise, updates are only performed once a week
here is my pushed tagshere is my packagistdoes anyone know how to resolve this?
Packagist does not take tags/releases that i push
Does your app use gunicorn?gunicorn will by default use synchronised workers and kill them after 30 seconds. On my Google App Engine app, it killed everything after 30 seconds regardless of if it's a cloud task or cron job.2 possible ways to get around it are:Use Async workers, see here:How can I run long tasks on Goog...
I am trying to schedule a task on app engine using cron that will run continuously on background. I have written the cron task it works fine on local server but when I run it through google cloud console it failed after 30 seconds.cron.yaml:cron: - description: daily tweets streaming url: /path/to/cron schedule: ev...
Google app engine cron django failed after 30 seconds
Some reasons: Building of Dockerfile could be long or computationally exhaustive and so building only once would be beneficial. Building of Dockerfile may require specific files/components that only exist on machine A or with person P and so giving a Dockerfile to someone to build just isn't possible without these so...
Closed. This question is opinion-based. It is not currently accepting answers. Want to improve this question? Update the question so it can be answered with facts and citations by editing this post. Closed 4 years ago. ...
Why should I use a docker image repository instead of building from a Dockerfile? [closed]
I realise this is 2 months old but I have been looking at this today and I found the existing answer not very useful.There doesn't appear to be support for full usage of regex, insteadonly the * and ? characters are used for regex matching. You can't use any of the characters outside of the supported range as shown bel...
Configuring an application load balancer.I have 3 separate target groups and about 30 different paths I'd like to filter on.When filling outpath patternwhat are the possible values? Given it's called "pattern" gives the impression it can accept regex but validation is preventing this.Is it possible to match multiple pa...
AWS application load balancer listener rule paths
amplify cli determines the status by diffing amplify/#current-cloud-backend and amplify/backend folder inside your project. So what you see when you run amplify status you see isn't accurate in your case. If you have created multiple environments (in different regions) make sure that you delete them too. The easiest w...
TL;DR: How to delete an amplify environment, when some resources of the service have been deleted manually in the console? So, I took a course on egghead to learn the aws amplify CLI. Unfortunately, it doesn't teach you how to delete the environment (otherwise it's great though!). My google search back then said you w...
AWS Amplify: How to delete the environment, when resources are already partially deleted?
I had a similar problem. In this case, my application used Basic Auth to access any link. So, I needed to add these lines to my configuration file :basic_auth: username: "username" # username with permission on the database password: "password" # password compatible with usernameAll credit toLucas Ribeiro B...
Please note: my prometheus is running using ubuntu terminal and my springboot application is running on windows. Seems like my ubuntu is not able to connect with the localhost of windows.I have created springboot metrics using "actuator" and my metrics are being exposed at "http/localhost:8080/actuator/prometheus".My ...
Prometheus target showing down
To solve the problem I opened the Visual Studio Installer and selected to update my Visual Studio 2017 Community installation.After the installation finished I started Visual Studio and opened the Team Explorer tab, selected Manage Connections and signed into Github again.After doing that everything worked again (like ...
I can't sync push or pull to my remote repository from VS2017. I am getting this error:Error encountered while pushing to the remote repository: Git failed with a fatal error. HttpRequestException encountered. An error occurred while sending the request. cannot spawn askpass: No such file or directory could not read U...
Visual Studio Github error
Yes.FromGithub API V3Documentation:For requests using Basic Authentication or OAuth, you can make up to 5,000 requests per hour. For unauthenticated requests, the rate limit allows you to make up to 60 requests per hour.For creating a gist, you can send aPOSTrequest as follows:POST /gistsHere's an example I made:<?php ...
By looking at GitHub Gist API, I understood that it is possible to create the Gist create for anonymous users without any API keys/authentication. Is it so?I could not find answers to following questions:Are there any restrictions (number of gists) to be created etc?Is there any example that I can post the code from a ...
How to create a GitHub Gist with API?
This should do the job for you:RewriteEngine On RewriteBase / # do not do anything for already existing files RewriteCond %{REQUEST_FILENAME} -f [OR] RewriteCond %{REQUEST_FILENAME} -l [OR] RewriteCond %{REQUEST_FILENAME} -d RewriteRule .+ - [L] # return /blank.php if no such file or folder in /fck/ RewriteCond ${REQ...
I have htaccess file that redirects all requests toindex.php, so that CMS system can parse the request and provide answer. However for obvious reasons this is not OK for files that actually exist on the requested path (in below case these are files uploaded by users using fck editor), therefore I have one of the rules ...
How to do in htaccess: if file in folder exists, serve it, if not, provide blank.html
I just ran into the same issue today.After some debugging I figured SES was instanciating using the wrong server (I'im using EU whereas US is default).ActionMailer::Base.add_delivery_method :ses, AWS::SES::Base, server: "email.eu-west-1.amazonaws.com", access_key_id: PLEASE_REMOVE_YOUR_CREDENTIALS_FROM_QUESTION, ...
I'm using Amazon AWS SES to send the common confirmation emails when a user get registered. I have my email and domain verified, but Rails doesn't send the message.I have installed aws-ses gem and it works because i've done some trys from Rails console. But when it has to send it automatically, i get:I, [2013-11-13T12:...
AWS SES and Rails. My app doesn't send the mails
Is it possible to destroy the queue in O(1)? Not if you intend to keep using a linked list. O(n) is optimal asymptotic complexity for destruction of a linked list (or any node based data structure). Also not if you store objects with non-trivial destructors. You cannot call N destructors in constant time. If you cho...
I implemented a queue data structure of my own. I implemented the queue using unidirectional linked list. And it is implemented in a class definition. The head of the linked list is pointing to the front of the queue and the tail of the linked list is pointing to the end of the queue. Note, I kept the tail only to mak...
How to destroy a queue efficiently?
1 This warning (No NetBeans projects added) can happen if the zip file was created manually. To avoid this create the ZIP file in NetBeans using File > Export Project > To Zip... Share Improve this answer Follow ...
I am trying to import a projetc (in ZIP file) in my netbeans. It is downloaded from GitHub (link: https://github.com/varavell/OnlineBookStore). While importing the project from Zip it shows the error and no project becomes imported. At the same time, at the bottom right corner it keeps unpacking of the project (never ...
NO NETBEANS PROJECTS ADDED warning
With proxy, when you site is up, all traffic will pass through that proxy. Now, what will you do when the proxy is down ? - While trying to handle one point of failure you just introduce an additional one. Also, you site response time will be lower, and you will pay three times for your traffic (your website, VPS in an...
I want to display beautiful page (with excuses) to the users if my webserver is down. How it is possible?My first idea was to make VM in cloud and to setup nginx there, which will check if webserver is available, and display beautiful error page if it's not.Is there another way to perform it (without nginx proxy)? (May...
Display special page if website server is down. Is there a way to do this without nginx proxy?
Seethis answer, you need to first create the object, and then assign a key with a variable by using themy_obj[x] = ysyntax.Your situationvar gist_description = "desc...", gist_content = "content..."; var data = { "description": gist_description, "public": true, "files": { } } var filename = "myfile.txt"; da...
I'm creating a github-gist using javascript/ajax.i want to provide my own file name. Below is the data which is in json.var data = { "description": gist_description, "public": true, "files": { "file.txt" : { "content": gist_content } } };i wrote var gist_filename = "main.txt";I replaced "file.tx...
using variable on left side of json syntax
If your IAM roles are setup correctly, then you need to download the file to the Sagemaker instance first and then work on it. Here's how:# Import roles import sagemaker role = sagemaker.get_execution_role() # Download file locally s3 = boto3.resource('s3') s3.Bucket(bucket).download_file('your_training_s3_file.rec', ...
I've uploaded my own Jupyter notebook to Sagemaker, and am trying to create an iterator for my training / validation data which is in S3, as follow:train = mx.io.ImageRecordIter( path_imgrec = ‘s3://bucket-name/train.rec’ …… )I receive the following exception:MXNetError: [04:33:32] src/io/s3_filesys.cc:...
Training data in S3 in AWS Sagemaker
Add the following to your Dockerfile: RUN apt-get update && apt-get -y install openssh-client
I am currently playing around with docker to create a PHP development environment. There are several containers (nginx, php7 + data container, mysql + data container, phpmyadmin) running and connected to each other, which is working fine so far. For convenience reasons, I am adding some scripts to the mysql container ...
Install SCP in Docker container
What you can do is only run the/gyanplease/rewrite for that host, like this (replacing the last two lines):RewriteCond %{HTTP_HOST} (?:^|\.)gyanplease.com$ RewriteCond %{REQUEST_URI} !^/gyanplease/ RewriteRule ^(.*)$ /gyanplease/$1 [L]That way this rule will only affectgyanplease.com.You can also change line 4 to this,...
I have 2 domains:gyanplease.com/gyanpleaseprimary domain (I have changed the document root with the help of.htaccess)mobiledevsolutions.comsecondary domainOn the document root I have created a folder calledgyanpleaseand with the help of.htaccessI have rewrittengyanplease.comto thegyanpleasethe folder. This the code for...
Redirect primary domain to sub folder without affecting secondary domain using .htaccess
This information is not sufficient to reproduce the problem. Please create a separate issue in the TestCafe Github repository using the followingformI would appreciate it if you provide us with your project or url to your site and your test code. It will help us determine the cause of the issue efficiently.
In a test, I click a link (SSL Cert) link that does not actually load the expected URL page. It hangs and as far as I can tell no issues when I look at network or console. Is there a known issue with TestCafe and clicking certain kinds of links. Clicking the same link with Selenium and Robot Framework successfully take...
TestCafe: Clicking SSL link does not load page
CauseFile.ReadAllBytes(fileName)also has to read the file which causes extra ~800 MBShareFollowansweredNov 8, 2012 at 14:36SnowbearSnowbear17.1k33 gold badges4444 silver badges6767 bronze badges11+1. Possibly of interest -stackoverflow.com/questions/4742016/…. I also would recommend adding some try/catch statements aro...
I am experimenting with Marshal.AllocHGlobal and found puzzling that this code would not succeed, instead it throws an OutOfMemory exception:namespace HAlloc { using System; using System.IO; using System.Runtime.InteropServices; class Program { static void Main(string[] args) { ...
Why would this throw OutOfMemory?
Well for the app itself it may not be so important, however having automated error logs can be very useful when you faced with an angry customer(s) and need to quickly debug the app. You can even have a special 'debug mode' to collect more info.You should also log your server transactions, adding an extra query in the ...
My experience are mostly in developing web applications and we do a lot of audit trails there. Literally every table is audited. I believe this is because user transactions are centralized to a server and they share the same table so it is important who did what.But now I am assigned to a project developing a standal...
How important is Audit Logging in a Standalone Application
SSH to the server and manually createuploadesfolder. You don't wantuploadesfolder to be in git. I assume it doesn't have any code, and all it has is user content - so just leave in on the server.This is how I push my local repo to a remote live server: I create a bare repo on the server, and push local repo to remote r...
okay i'll explain my issue short:I have a website, lets say a board where user can chat and upload avatars, images, files etc.I want to use git with this website.So what i did:created a repo with bitbucketclone this repo localadding a .gitignore file (see below)copied my website to my local repomade a first commitpush ...
Using git with a website that has user uploads
The document is for WebSphere Application ServerNetwork Deployment, not BASE edition. This particular admin section is used to manage certificates usedbetweenthe components within the cell. Base does not offer cells, nodes, clusters, ...
I need to set-up the Endpoint security configuration (Reference - This IBM Doc)When I do Admin Console -> Security -> SSL certificate and key management -> Manage endpoint security configurationsAccording to the IBM doc, I should be seeing Inboind / Outbound links, but I don't see any links that I can set up for certif...
Websphere BASE version - Manage endpoint security configurations
I assume that inputdata is not NULL.In that case the second approach should not work at all, since the specifications says, that clCreateBuffer returns NULL and an error, if:CL_INVALID_HOST_PTR if host_ptr is NULL and CL_MEM_USE_HOST_PTR or CL_MEM_COPY_HOST_PTR are set in flags or if host_ptr is not NULL but CL_MEM_CO...
I have seen both versions in tutorials, but I could not find out, what their advantages and disadvantages are. Which one is the proper one?cl_mem input = clCreateBuffer(context,CL_MEM_READ_ONLY,sizeof(float) * DATA_SIZE, NULL, NULL); clEnqueueWriteBuffer(command_queue, input, CL_TRUE, 0, sizeof(float) * DATA_SIZE, inpu...
What is the difference between creating a buffer object with clCreateBuffer + CL_MEM_COPY_HOST_PTR vs. clCreateBuffer + clEnqueueWriteBuffer?
Please check your network on the guest-mashine with: nmap -sT -O localhost Check if the ports your are using in your nginx configuration are open. If not, open them in your firewall and check again.
I have set up a Vagrant machine with this configuration -- Vagrant.configure("2") do |config| config.vm.box = "intprog/centos7-ez6" config.ssh.insert_key = false config.vm.network "public_network", ip: "192.168.33.243" config.vm.provision "file", source: "/server/bin/nginx/conf/domains-enabled/cemcloudM...
Vagrant - centos networking
After a friend told me how roughly the filter caches works, it become clear why we get out of memory errors from time to time. So what does the filter cache do? Basically it creates something like a bit array which tell which documents matched the filter. Some something like: cache = [1, 0, 0, 1, .. 0] 1 means it hit...
I have a Solr setup. One master and two slaves for replication. We have about 70 Millions documents in index. The slaves have 16 GBs of RAM. 10GBs for the OS and HD, 6GBs for Solr. But from time to time, the slaves are out of memory. When we downloaded the dump file just before it was out of memory, we could see that ...
Solr Filter Cache (FastLRUCache) takes too much memory and results in out of memory?
First you have to change save file extension .caf to .wav in below line given[self.voiceHud startForFilePath:[NSString stringWithFormat:@"%@/Documents/MySound.wav", NSHomeDirectory()]];Then goto the method (startForFilePath:) change the settings for the voice quality.NSDictionary *recordSettings = [NSDictionary ...
I am working on audio recording. I am able to record my audio in .caf (Core audio format)and later I need to convert it to .wav or .wma in order to upload the file on FTP Server. How can I convert the file to .wav or .wma format in iOS? Using POVoiceHUD code from GitHub. Can any one help me to solve my problem.
How to convert .caf file to .wav or .wma file using this code POVoiceHUD on GitHub
TL;DR - use name instead of Name as the key. To investigate, I turned to my local installation of ec2 tools, and ran ec2-describe-images -o self -F name=myaminame Got a similar error that was more Google friendly: Filter definitions must have format 'name=value', but found 'name' Googlging got me to this blog post,...
I'm trying to check whether a specific AMI exists. So, I'm doing: val filter = new Filter().withName("Name").withValues(amiName) val result = ec2.describeImages(new DescribeImagesRequest().withFilters(filter)) result.getImages.size() > 0 (code is Scala and not Java, but that's not really relevant). I'm getting the fo...
How to filter described images by name in DescribeImagesRequest?
1 No, you can't. What you could do, is setting up WAL archiving to make incremental backups: http://www.postgresql.org/docs/current/static/continuous-archiving.html#BACKUP-ARCHIVING-WAL This can only be done for the whole cluster, not for a single database. Share ...
I have a decent sized PostgreSQL database (approx 6GB & growing). A full backup/export of the database is done every few hours via cron & pg_dump. Specifically, can I export only the changes to the database since the last export? Or perhaps run a utility that compares the two exports and appends the differences to t...
Export only new data since the last PostgreSQL database export
You could:Create the PR when the work is ready and not before.Work with the 2 protected branches workflow, e.g.develop->main(ormain->release/live/etc.) where you first put features intodevelopand only when desired you merge them intomainwhich has the Jenkins task attached.You could not run the check automatically on pu...
I have a protected brach in github with require status check - Jenkins build.When a user is creating a pull request, The Jenkins test is starting and that's ok. But when a user is pushhing commit to the branch of the pull request - the test is starting again, even if the user didn't finish to work on this branch, and h...
github pull request ignore status checks after pushing commits
You need to pass thestatusCodeafter executing the Lambda function. If you don't pass it the API Gateway will trigger502 Bad Gatewayby default.message = { 'message': 'Execution started successfully!' } return { 'statusCode': 200, 'headers': {'Content-Type': 'application/json'}, 'body': json.dumps(message)...
First of all, I am new to whole AWS itself. I am trying to solve a problem using AWS lambda and API Gateway.I created a python lambda function. It takes some data and returns astringbased on some conditions. Something like below:def lambda_function(event, context): if event['some_property']: return "SUCCESS...
AWS Lambda and Gateway API Integration, returns status code 500
it would seem better to flip things around and put the upward growing part (text, global data, and heap) on top? No. Cache lines are typically only 32 bytes to 256 bytes. It's rare for a program to use less than a few megabytes of data, so the sharing is basically irrelevant. (Even if you're not using it, the standa...
I've read that on linux, program memory layout can be broadly visualized as follows (and I assume it's similar on most other operating systems): Now, I'm not sure if I remember correctly, but I think CPUs read fairly large chunks of adjacent RAM addresses into their largest cache level in a single go. If so, it would...
Does the distance between a program's stack and data segment have an effect on CPU caching?
Put this code in yourDOCUMENT_ROOT/.htaccessfile:RewriteEngine On RewriteCond %{REQUEST_URI} !\.html$ [NC] RewriteRule ^(.+?)/?$ /$1.html [L,R]ShareFolloweditedNov 14, 2013 at 10:17answeredNov 14, 2013 at 10:03anubhavaanubhava771k6666 gold badges582582 silver badges649649 bronze badges4i tryed this one, the /someurl i...
I'm working with html5 pushState(), but if i wanna load the site by a pushState URL it returns 404.Therefore i need to rewrite all /somePushStateURL to /somePushStateURL.html to get to the right site.How the right htaccess entry will look like ? Are there some other solutions for my problem ?
redirecting url for pushstate support to .html url
3 The SageMaker inference endpoints currently rely on an interface based on Docker images. At the base level, you can set up a Docker image that runs a web server and responds to the endpoints on the ports that AWS require. This guide will show you how to do it: https://d...
I have a custom machine learning predictive model. I also have a user defined Estimator class that uses Optuna for hyperparameter tuning. I need to deploy this model to SageMaker so as to invoke it from a lambda function. I'm facing trouble in the process of creating a container for the model and the Estimator. I am a...
How to deploy a custom model in AWS SageMaker?
The best way to add tags is to use the Prometheus service discovery. This keeps these tags out of your application code and keeps it from being concerned about where it exists.However sometime if you absolutely need those extra tags (due to the service having extra insight that Prometheus service discovery isn't surfac...
I have few spring-boot microservices with actuator and exposed prometheus metrics. For example:# HELP process_uptime_seconds The uptime of the Java virtual machine # TYPE process_uptime_seconds gauge process_uptime_seconds 3074.971But there is noapplicationtag, so I'm not able to bind it to a certain application within...
Is there any way to add `application` tag to standard prometheus micrometer metrics?
The reason you keep running afoul of the invalid certifcate error is due to howHTTPSworks. Namely, the HTTP connection is tunneled over TLS. This means that the client initiates the TLS connection to the web server (including the process of verifying the server certificate) before getting to the HTTP layer, where your ...
I'm trying to redirect the users fromhttps://www.example.comtohttps://example.com(because SSL cert is valid for the later only), and this works great in Chrome with this code:if(isset($_SERVER['HTTPS']) and $_SERVER['SERVER_PORT'] == 443){ // We are on https version if(strtolower($_SERVER['SERVER_NAME']) == 'ww...
Firefox won't redirect from HTTP to HTTPS properly?
Depending on the Image you are using for kafka you can supply these settings via the environment variableKAFKA_OPTS.The documentation you are referring to is supplying these options to the call of'java'. Kafka, Zookeeper etc. are jars and there for stated via java.
I want to set the Max and minimum memory value for the Kafka server [9092 port]Let say Max value is 2 GB, then memory usage should not exceeds the 2GB, but currently exceeds it.I have link -https://kafka.apache.org/documentation/#javaConfig From Apache site-Xmx6g -Xms6g -XX:MetaspaceSize=96m -XX:+UseG1GC -XX:MaxGCPause...
How to Limit the kafka server Memory?
To apply the headers to all request, you need to write a middleware function(server-side):// set header function function setNoCacheHeaders(ctx) { ctx.set('Cache-Control', 'no-store, no-cache, must-revalidate') ctx.set('Pragma', 'no-cache') ctx.set('Expires', 0) } // Middleware that adds the header to all reques...
I am working with an app built with KOA framework and I'm trying to figure out why a page is cached. In all browsers even a hard reload won't work. You literally have to clear cache to see the page update.I want to add this to my index.js but I do not know where to add the line.Can anyone help?ctx.set('Cache-Control', ...
Set header cache control in KOA framework
Have you tried this ?First, you need to open HTTPS port (443). To do that, you go tohttps://console.aws.amazon.com/ec2/and click on the Security Groups link on the left, then create a new security group with also HTTPS available. Then, just update the security group of a running instance or create a new instance using ...
I have an application running on an AWS EC2 instance with the domain's nameservers on AWS as well. I have an A record with the public IP.I've create a secure certificate with ACM and also created an ELB Load Balancer. My domain still doesn't show the HTTPS in front of it.Can anyone provide some help? Many thanks
Adding a secure HTTPS certificate to AWS EC2 Instance
I'm sorry, I cannot speak to anythinghelmbased. I can speak for a concourse pipeline, though.Concourse does not support providingparamsto tasks that are an array.paramsare passed in as environment variables to a running task, so they are transformed from YAML to a simple string key-value pair.If you want to pass more c...
I'm trying to figure out how to create an array with some CIDR ip address in order to have the same array in my pipeline. So here is an example var file:whitelist-ip-ranges: |- - 10.0.0.0/24 - 11.0.0.0/24 - 12.0.0.0/24My pipeline is like:.... .... .... params: variables: loadBalancerSourceRanges: ...
Concourse CI - array variable
I was missing a label for the east replicaSets, after i add the app:consumerAPP, it works fine now. Thanks
we have a system that is having 2 endpoint based on geo-location. e.g (east_url, west_url). One of our application need to load balance between those 2 urls. In the consumer application, created 2 deployment with the same image but different environment variables such as "url=east_url", "url=west_url".after the deploym...
K8s loadbalance between different deployment/replicasets
The command isopenssl verify -CAfile ca.crt server.crtIf the two certificates match, the command will returnserver.crt: OKThe above command is only for pem format.If you has .p12 file, then first convert it to pem format:openssl pkcs12 -in server.p12 -out server.crt -nodes
I have a SSL server certificate for a server and I have a root CA certificate that the client can verify this server certificate during SSL handshake. For some reason I am not sure if I really retrieved the right root certificate. Is there any way where I can see if I have the right root certificate? Maybe an openssl c...
verify if server certificate matches with root CA certificate
You should use thetemplatingfeature of the Zabbix Grafana plugin,see the attached screens and the following description for a working example.I have a Routers Zabbix Hostgroup, so I define aRouterGrafana variable to match the hosts (Routers.*), see the first screenshot. Enable both multi-value and Select AllThen in the...
I want to show multiple CPU Usage from different hosts in one graph but they all end up with the same name and I can't tell which line represents which host:here's the snapshot.I'm using Grafana 5.2.4 with a Zabbix plugin 3.9.1. My Zabbix version is 3.0.12. I've tried overriding legends in Grafana but there's no such o...
Can't Change Metric Alias in Grafana Using a Zabbix Plugin
No unfortunately there is no better way to retrieve commits for the user. However there is a workaround for the rate limit:DocumentationsaysFor requests using Basic Authentication or OAuth,you can make up to 5,000 requests per hour. For unauthenticated requests, the rate limit allows you to make up to 60 requests per h...
Is there an efficient way to find all recent commits by a specific user across all public repos?I am currently use/events/publicand filtering out thoseevent.type === "PushEvent". However this is not very efficient becauseThe commits in thePushEventdoes not havetimestamp, which means I need additional requests to fetch ...
GitHub API find all commits by user ID with timestamp
0 I think you are using modal segue to go back to the previous scene. Then , when every segue is performing there are new allocations ... Just use this method to go back previous scene: [self dismissViewControllerAnimated:YES completion:nil]; and delete the second moda...
I have a problem with growing memory using StoryBoard. I use modal segues to get to the next or prev. Viewcontroller. Sample UIStoryBoard When i run Instruments i can see the Viewcontrollers never gets out of the memory. So they are added and added each segue. Instruments Allocations How can i get this Viewcontroller ...
Abandoned memory after Modal segue with StoryBoard and ARC
Well your site root is a directory, so this rule you have excludes existing directories. What you could do is only exclude existing files, and allow existing directories to be handled by the PHP script. Like this:RewriteEngine on RewriteCond %{REQUEST_URI} !=/api/init.php RewriteCond %{REQUEST_FILENAME} !-f RewriteRule...
I have checked a large amount of existing answers regarding.htaccessredirects. However none of them have helped me.What I want to accomplish is redirecting all request urls to/api/init.php. However I've only gotten so far to where my index pagewww.example.comsimply gives me a file listing because of the missingindex.ph...
Redirecting all urls, including no path, to a file in subdirectory
I used req.get_uri_args() to get all parameters passed in url.local args = ngx.req.get_uri_args()ShareFollowansweredNov 15, 2013 at 10:34Prashant GaurPrashant Gaur9,7001111 gold badges5151 silver badges7171 bronze badges3It didn't work for me.. it gives 500 internal server error. Can you share how you did it in more d...
I am having a url '/gifts/' below is code of nginx.conf file which is managing logic.location /gifts { default_type text/html; set $target ''; content_by_lua ' local redis = require "resty.redis"; local red = redis:new() red:set_timeout(1000) -- 1 sec local ok, err = red:con...
fetch parameters passed with url into nginx.conf by lua in Openresty
I was having this same issue and determined it to be coming from the following line in ebcli/operations/deployops.py:io.log_info('Deploying code to ' + env_name + " in region " + region_name)I resolved it by explicitly specifying a region with-r REGIONwhen executing the deploy. I imagine somewhere the default region sh...
Sorry I don't see this question anywhere. I've deployed an environment for my application "scones" on AWS. It's fairly simple Node.js express app. Honestly it's just to test the elastic beanstalk deployment. Anyhoo I'm using ebcli and I've already done an "eb create scones1" and it does create the environment and launc...
eb deploy ERROR: TypeError - cannot concatenate 'str' and 'NoneType' objects
Yes, you can start additional Java process on new port 8081 and change firewall rule to that port. Whether the connections will be dropped in case of restart is determined by option IPTABLES_MODULES_UNLOAD. If this option is set to yes (default) - they will be dropped, if no - then they should not. I would suggest usin...
If enable this ruleiptables -t nat -A PREROUTING -p tcp --dport 80 -j REDIRECT --to 8080Then connections coming to the server port 80 are redirected to localhost:8080. If I want to restart the service, can I just start the service on another port? Say port 8081 and re route the firewalling toiptables -t nat -A PREROUT...
How to restart a Java service without disruption using firewall/iptables?
I feel like I have to give this a proper answer because coming upon this question was the answer to 10 straight hours of debugging for me. As @Daniel said in his comment, the issue I found was with my ACL blocking outbound traffic on UDP port 53 which apparently kubernetes uses to resolve DNS records.The process was e...
I have an EKS cluster setup in a VPC. The worker nodes are launched in private subnets. I can successfully deploy pods and services.However, I'm not able to perform DNS resolution from within the pods. (It works fine on the worker nodes, outside the container.)Troubleshooting usinghttps://kubernetes.io/docs/tasks/admin...
DNS problem on AWS EKS when running in private subnets
Zabbix can be a choice:https://www.zabbix.com/integrations/activemqIf the queue size becomes huge, it will create an alert. Also, it has more variety of other alerts that can be useful.
Hi with activemq problems that i usually face are as under:- 1. Pushed message do not get consumed 2. As a result of point 1, queue size becomes huge 3. Messages in queue keep on retryingSo is there any open source that anyone of you have used in project to send mails when queue size goes beyond a certain threshold. Ca...
Monitoring and alerting on activemq
drawable can be pushed to GitHub as follows:In the left panel of Android studio, right click your drawable directory.Choose git, add.(Ctrl+Alt+A)Following steps are as usual.Done!ShareFollowansweredJan 24, 2016 at 18:03W.ManW.Man66588 silver badges1212 bronze badgesAdd a comment|
I want to upload my codes to GitHub, so I created a repository and uploaded the code initially. But the problem goes after that, when I try to push, Android Studio is telling that everything is done, but when I go to web, I see, everything is as like it was.Let me share my screen shots here-1-Starting commit2 - Commit ...
Android Studio - GitHub Push not working perfectly
You can first pass the secret to an env var, then echo it to the file, remember to quote the env var to prevent line feed from breaking the command.jobs: build: runs-on: ubuntu-latest steps: - name: print secrets env: MY_SECRET: ${{secrets.KEY}} shell: bash run: | echo...
I was trying to do the following :echo ${{secrets.key}} > myfileBut unfortunately, this doesn't work since myfile would be empty after this when i checked. How do i save the content of github secret into a file ?
How to save github secrets content in a file in github action
You should not reinitialize your git again for a separate project, go to your git directory, and clone the new project usinggit clone https://github.com/ricke2005/Learn-rails-2.gitthen move to this new project directoryLearn-rails-2in your git folder, and you can start working on that project.
When trying to add a new GitHub repository to my project, I am gettingfatal: remote origin already exists.Then, when I try to dogit push -u origin masterI get another error and it gives an error of faulty path. This path continues to show up even after I typed it in correctly like it keeps reverting back to my original...
GitHub initialization trouble
As per thegithub doocumentfrom 15th march github will permanently stop acceptinggit://protocol. You can change the url fromgit://tohttps://.Refer thisgithub issuefor more info
Retryinggit clone 'git://github.com/Azdaroth/simple-captcha.git' "/home/intrcity/.rvm/gems/ruby-2.5.0@ry_rails5/cache/bundler/git/simple-captcha-f06f0cfd255417fffe861c65e1badbe22e822021" --bare --no-hardlinks --quietdue to error (2/4): Bundler::Source::Git::GitCommandError Git error: commandgit clone 'git://github.com/...
I am getting error on bundle install in RubyOnRails
It looks like you are trying to run your script using a directory "/usr/local/bin". Presumably you either want to run the script using python:python -q /home/bdmweath/public_html/scripts/my_script.pyOr make the script executable and run it directly:chmod +x my_script.py ... /home/bdmweath/public_html/scripts/my_script....
I'm new to the whole website world, so I apologize before hand if this is a duplicate question of some kind.Important note: I am well aware this specific script won't update the graph. It's just a representation of the script's file path and the output that I want. The graphwill updatewhen my script is in place and the...
Cron Job - Python matplotlib script
Try below possible Solutions : I)Memory Issue : 1)Try adding more resources (requests & limits) under the container - may need more memory. 2)Try to increase the available amount of RAM that the docker container, Please see SO for more info and also Clear the Cache : log in the console of the backend container, then i...
So basically i am trying to deploy a service in Kubernetes cluster using a DockerFile, after so many attempts i am still stuck with the same error message Booting worker with pid: 7 on loop, without any exit code i.e. [2023-01-11 09:22:48 +0000] [1] [INFO] Starting gunicorn 20.1.0 [2023-01-11 09:22:48 +0000] [1] [INFO...
Stuck on "Booting worker with pid: 7" while deploying a service in Kubernetes using Docker image, What seems to be the problem here?
To run cron every 5 minutes you need to add command like this:*/5 * * * * /home/cdh_infa_user/data/pladmin/MyLinuxAgent/apps/Data_Integration_Server/data/scripts/Secureagent.shTo run cron at 5 a clock you need record like this:0 5 * * * /home/cdh_infa_user/data/pladmin/MyLinuxAgent/apps/Data_Integration_Server/data/scr...
I am trying to run a crontab with the expression given below. But i am getting bad minute error.This is for a Linux Server.0/5 * * * * /home/cdh_infa_user/data/pladmin/MyLinuxAgent/apps/Data_Integration_Server/data/scripts/Secureagent.shDo i need to install crontab? Please guideenter image description hereenter image ...
Error in cron: bad minute errors in crontab file, can't install
From Commentstf.test.is_gpu_available()tells whether Tensorflow can access aGPU.THIS FUNCTIONIS DEPRECATED. It will be removed in a future version. Instructions for updating: Usetf.config.list_physical_devices('GPU')insteadIf it returnsTrue, then there is no issue in using GPU. ButGeForce MX110is slowest since it also ...
I have trained a faster R-CNN model for object detection using TensorFlow object detection with Google colab. But when testing videos google colab crashes, that's why I decided to test on my pc and installed CUDA 10.0 and Cudnn 7.6.5 and TensorFlow-gpu = 1.15. But the test is so so slow as if it is running on a CPU. I ...
Using GPU for tensorflow object detection
There's a class for that;MSDN Article on System.Net.NetworkInformation.PingGeekPedia article on asynchronous ping
In Code i want to validate a domain name.For example : " DomainName.com".How can i do that in C#.I worked onMSDNSolution. (Second Solution).But"PingCompletedCallback" is not getting executed.Thanx
How to "ping" a domain name?
The approach to create a pipeline in a runspace e.g.:var pipeline = runspace.CreatePipeline();is 1.0 thing. That is, the original PowerShell hosting API required you to create the pipeline through the runspace that you created. My guess is that the team got feedback that the hosting API needed to be simplified so the...
I am using powershell commands to execute scripts and cmdlets. So while executing cmdlets I used powershell.invoke and while executing a script I used pipeline.invoke method. I wanted to know if there is any difference between theSystem.Management.Automation.pipeline.invoke()method andSystem.Management.Automation.Runsp...
What is the difference between pipeline.invoke and powershell.invoke?
You can usehttps://wordpress.org/plugins/wp-crontrol/Cron control plugin which is very useful to add and manage all the cron jobs.ShareFollowansweredJul 19, 2016 at 12:13KappaKappa1,01311 gold badge1616 silver badges3131 bronze badges2This plugin don't give which plugin is reponsible for scheduling the cron job. And th...
I'm running a Wordpress blog on a512MB VPSand due to the lack of memory,MySQLcrashes frequently.I'm trying to optimize the server and I need to know what are the plugins that use cron jobs. I found the Wordpress method_get_cron_array()that returns an array of all currently scheduled tasks but I don't have any informati...
Wordpress: List of cron jobs by plugin
this is too old a question but for this, it has already several answers in StackOverflow, in that one of the best and I used methods is using the node-openvpn package.first thing first we need to install the packagenpm install node-openvpnat your index.js or server.js use below codeconst openvpnmanager = require('node-...
I am a newbie to node.js and looking for some example code or pointers on connecting through PPTP to a private virtual ip address using VPN connection. I have a node.js server running on aws that currently uses udp to connect to a public ip address. However, this needs to be changed to tunnel into the private vpn.I hav...
How do you connect to vpn using node.js
Thanks for the question! You can set any HTTP header in an HTTP response. For instance: onRequest(HttpRequest request, HttpResponse response) { ... response.headers.add("Cache-Control", "max-age=3600"); ... } If you want more sophisticated handling, such as respecting Etags or If-Modified-Since, you'll probably...
So, I've noticed that using Dart's built in HttpServer class tends to make the client request for every file every time. On Apache, it is possible to tell the client to cache the file for a maximum of a certain length of time -- does Dart support this feature to lighten the load on HttpSever?
Dart's HttpServer and client-side caching
rpivotTableis sort of a wrapper around a js librarypivotTable.Hereyou can find all its parameters. Try using the one calledmenuLimit, for example:tmp <- rbind(iris, iris) # Here you have limit on num column: rpivotTable(cbind(tmp, num = 1:dim(tmp)[1])) # Here you don't: rpivotTable(cbind(tmp, num = 1:dim(tmp)[1]), menu...
I have created an rpivottable and my problem with it is that when it runs, I can't click on one of the options on the side bar to filter out a specific factor because it says there are too many to list.Pivottable:And then when I click on Ad.Source, this comes up:This is a problem, because if I want to create line graph...
RPivotTable being able to list more than allowable amount
boto3.resource is a high-level services class wrap around boto3.client. It is meant to attach connected resources under where you can later use other resources without specifying the original resource-id. import boto3 s3 = boto3.resource("s3") bucket = s3.Bucket('mybucket') # now bucket is "attached" the S3 bucket ...
I am trying to understand when I should use a Resource and when I should use a Client. The definitions provided in boto3 docs don't really make it clear when it is preferable to use one or the other.
When to use a boto3 client and when to use a boto3 resource?
You need to write two dashes instead a single one:... --log-driver=gelf --log-opt ...EDIT: In that blogpost it is written:The Docker plugins mechanism is now available in the new Dockerexperimental channel.You need to install the experimental docker version like:curl -sSL https://experimental.docker.com/ | sh
I've been installing the last version of Docker (1.8). This new version is great because you can set a particular log-driver to send the log to Graylog2 (for example).Here's the blog post from docker :Docker Blog announcing 1.8Here's the graylog team blog to make the magic happen :Graylog blogMy problem is that it doe...
Docker & Graylog
/** * Get directory path where backups stored * * @return string */ public function getBackupsDir() { return Mage::getBaseDir('var') . DS . 'backups'; } is the function in class Mage_Backup_Helper_Data. You need to log value of whats returned by this function by changing code to ...
i've this strange behaviour on my magento installation: every time i try to launch a backup, the file is saved into var folder instead of var/backups folder, so that is not visible into backups list (whose looking for var/backups folder). Any suggestion? Nothing has been changed since yesterday. Thanks
Magento backup folder
Since SSRS and SSIS are XML files, you could compare the text with the source code. It sounds like you might need to create your own custom code to compare.Some questions I would have when doing something like this:Are the production files in a similar folder structure to the source code?Is there a standard naming con...
The company I work for has recently made a new IT division. With this new IT division we wanted to do a full audit of all the current code we have out in our source control but with a few thousand projects for Business Intelligence alone manual process of looking at the source control code and comparing it to productio...
Any good ways of quickly auditing code in a source control compared to production?
Your link comes from the service shields.io, which offers badge images for all sorts of services and all sorts of info. The reason you had a hard time finding documentation is probably because badges aren't GitHub specific, that's just one place you can use them. So if you search for "nuget badge" instead of "nuget gi...
I have a repository on GitHub and I've created a corresponding packing on NuGet. I would like to link to the NuGet package from my GitHub README file. After modifying some text I found on other repositories, I came up with the following: [![NuGet version (SoftCircuits.Silk)](https://img.shields.io/nuget/v/SoftCircuits...
Linking to NuGet from GitHub
It's not compatible with version 3.7. Seehttp://docs.codehaus.org/display/SONAR/Plugin+version+matrix
I'm looking into upgrading Sonar, from 3.2.1 to 3.7.1. In our current system, we have the fb-contrib plugin installed and many of its violations enabled. However, in the latest version, I do not see it listed in Update Center -> Available Plugins. I've looked at the release notes for each release and I do not seen any ...
Does Sonar 3.7.1 still support the fb-contrib plugin?
+200You can use the above schema. Now for you queriesTopics by Forum (sorted by date posted, or most recent reply)Select from GSI1 where GSI1 pk=Forum123 and sortby GSI1 SKyou can choose whom to keep in GSI1 Sk recent reply/date posted based on which use case is frequently asked.Replies by Topic (sorted by date posted ...
The AWS DynamoDB documentation includes an exampleschema for a forum. However, the number of questions this schema is able to answer seems very small. In addition, the table seems to suffer from a hot-key problem (a burst of replies backs up on the same partition).In a talk title "Advanced Design Patterns for Amazon Dy...
How to model a forum using Amazon DynamoDB without hot-partitions?
Each32-bitpointer in Java allocates ~4 bytes. Hence, it will be5736 * 4 = 22944 bytes
In Java, on a 32 bit OS, I store5736 int[]arrays containing54759118ints. If eachintprimitive takes4bytes, then I will get minimum54759118 * 4 = 219036472 bytes ~ 220 megabytes. Since I have5736references toint[]objects, what will be an overhead of storing those references?
Memory size of Java 32-bit system int[] arrays
Assuming you are usingboost::shared_ptr,std::shared_ptr, or something else that is similar, then whenAis destroyed, the shared_ptrs will all have their reference counts decremented by one, which may cause their own destruction.
I have an object that has a map of shared pointers.class A { map<string, shared_ptr<B>> child; }Would the destructor for A take care of destroying the pointers in the map? Do need to manually reset the map's values?I am seeing a _Resetp<> and _Ref_count hold on to 2 objects and I was curious if the problem was that ...
What is the correct way to clean all the shared pointers that are in a map?
15 Amazon's Route 53 Health Check is the right tool for the job. Route 53 can monitor the health and performance of your application as well as your web servers and other resources. You can set up HTTP resource checks in Route 53 that will trigger an e-mail notification...
I've got an app running on AWS. How do I set up Amazon CloudWatch to notify me when the EC2 instance fails or is no longer responsive? I went through the CloudWatch screens, and it appears that you can monitor certain statistics, like CPU or disk utilization, but I didn't see a way to monitor an event like "the instan...
How do I set up CloudWatch to detect when an EC2 instance goes down?
No, you can't easily do this with just one cronjob.You could, if you combine cron with some other scheduler, as in0 * * * * otherScheduler && /cronjob.sh. But that's more complicated than simply two cronjobs.You also could do this by changing the date on your server such that the cron schedule is from February 15 to 28...
Is there any way to schedule the CRON job for different days of different months!For Example: I need to schedule the job from February 25 to March 10 2017.this can be done by creating 2 jobs as "0/1 * 25-28 2 2017 /cronjob.sh" and "0/1 * 1-10 3 2017 /cronjob.sh"But, cant I do it in one job? is there any way to do tha...
Cron Job Scheduling Different Dates of different month
When CppDepend is added as an extension of SonarQube, you have to activate all its rules in your quality profile. For that you can follow these steps:Create a new sonarqube C++ quality profile.Extend this quality profile to add the CppDepend rules.Make this new quality profile the default quality profile.
I created a new rule on cppdepend. My cdproj document included the new line as follows<Query Active="True" DisplayList="True" DisplayStat="True" DisplaySelectionView="False" IsCriticalRule="False"><![CDATA[// <Name>Classes must start with an C</Name>How can I make it appear on the sonarQube dashboard as code smell?
How can I define a cppdepend rule as code smell rule?
To get a shell in a running container, do this:$ docker exec -it <container-id> /bin/shTheattach sub-commandgives you access to a running containers stdout. That's not what you want hereHowever, if your conainer is meant toprovideSSH as a service, you'll need torun it in such a waythat it's exposed on the host, on some...
I am following this blog on how to connect to a docker instance:https://phoenixnap.com/kb/how-to-ssh-into-docker-container. It mentions usingdocker attach <name>Trying this on my ec2 instance gives us:$ docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS...
Unable to ssh to Docker instance using `docker attach <name>`
If you don't cut your XML output Length itself i don't see a possebility for a big speedup.Stripping unnecessary whitespaces for transport is pretty usefull because it could reduce transport speed. even if it is not that mutch. the longer your xml is and the more unnecessary whitespaces you have there (unnecessary for ...
HI,I have a cron that calls a basic API URL. The output is xml so I use php'sfile_get_contents()to get the data to process.My question is, Does the output format of the xml make a difference in the transfer time from one server to another? The cron is running every ten minutes and I don't want the crons to overlap at s...
How can I optimize my xml output so that the data transfer time be lower?
This should do what you want:Options +FollowSymLinks RewriteEngine On RewriteBase /phone/ RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^(.*?)/?$ index.php?menu=$1 [L] RewriteRule ^(.*?)/(.*?)/?$ index.php?menu=$1&submenu=$2 [L]//OR the following if the values are always going t...
I have two URL samples in my new site.sample1:http://localhost/phone/3sample2:http://localhost/phone/3/10please note: phone is the project nameAnd I wrote an htaccess file myself, but it's only working with sample1.Options +FollowSymLinks RewriteEngine On #RewriteBase / RewriteCond %{REQUEST_FILENAME} !-f RewriteCond...
.htaccess is not working on all links
This is because your config app.config['SERVER_NAME'] = '0.0.0.0:5000' Your Flask thinks its name is 0.0.0.0:5000, therefore not responding to others. To verify, call curl -H Host:0.0.0.0:5000 -i http://app:5000 from test. You can remove this config and run the server with app.run(host='0.0.0.0')
yet another docker/flask question. I'm having problems connecting to my flask app from another container in docker-compose. My dockerfile sets up a flask app like: COPY ./ /app/ WORKDIR /app RUN find -name "*.pyc" -delete && \ pip3 install -r requirements.txt EXPOSE 5000 CMD ["python3", "app.py"] The app runs...
Problem reaching flask app in docker-compose
No you should not commit this file. The wholeLibraryfolder isn't supposed to go into the repository, because it's generated by Unity.What you need to do is: Add a.gitignorefile to your repository and commit it (as a single file). From then on git will ignore theLibraryfolder and a lot of other things. It will keep work...
I've created a new, currently empty Unity project (Unity version 2021.3.15f1) and have created a Github repository for it. I was having trouble with the initial commit--it kept throwing an error at me--so I tried following this guidehttps://unityatscale.com/unity-version-control-guide/how-to-setup-unity-project-on-gith...
Initial Github Repository Commit For Unity Project, One Filename Too Long
No, you can't do that.For many use cases, there are workarounds. For mine, there isn't.
Given a time series, is it possible to find the timestamp of the last value change?For example, a cron job runs every night. It first runs an rsync job, and later computes the size of the target directory. The actual metric value is not all that significant, but I would like to verify that it actually changes every nig...
Can I find the timestamp of the last value change
Note that you never "connect" to github.com directly (no interactive shell)Asmentioned here, "Connection refused" meansinvalid IP address forgithub.com(manual entry in/etc/hostsor your resolver)firewall along the way to github.com which blocks the ssh traffic (eg. local firewall or corporate firewall)So is the other ma...
Essentially I cannot connect to an SSH server anymore. Whenever I try to connect I get the following error:ssh: connect to host HOSTNAME port 22: Connection refusedNote that I am able to connect from a different machine. Hence, the issue seems to be client side. Furthermore, I also cannot pull/push to git repositories ...
Solving the a client side SSH connection error
I found a way to restart nginx after deployment using an undocumented technique for running post-deployment scripts. I added this to my .ebextensions:files: "/opt/elasticbeanstalk/hooks/appdeploy/post/03_restart_nginx.sh": mode: "000755" owner: root group: root content: | #!/usr/bin/env bash ...
I'm running a rails application on Ruby 2.0/Puma instances and am trying to customize the nginx configuration. I need to increase the permitted request size to allow file uploads. I've found some other posts that have lead me to add this to my .ebextensions:files: "/etc/nginx/conf.d/proxy.conf" : mode: "000755" ...
Customizing Nginx Configuration in AWS Elastic Beanstalk
Explicitly calling process.exit forcibly discards some pending asynchronous tasks. Calling process.exit() will force the process to exit as quickly as possible even if there are still asynchronous operations pending that have not yet completed fully, including I/O operations to process.stdout and process.stderr. -...
I wrote a code that ends the process under certain conditions, and when I write it as a return, it ends abnormally due to an over-memory error, and process.exit() ends normally. P.S. My code is just only one function. So the two method quit the one function. Can you explain the difference between the two methods when...
What is the difference between process.exit() vs return when ending the process?
You can do like:// Set the environment variable, NAME. putenv('NAME="Robert Smith"'); // Retrieve the environment variable, NAME. echo getenv('NAME'); // Unset the environment variable, NAME. putenv('NAME');Setting environment variables in an Apache configuration If you're using Apache as your web server, then you ca...
I want to store my credentials in a secure place and can use them as env variable in my file in cpanel. I don't want to use any library or composer method to do this. I have seen solutions for the same on web by using .htaccess file but didn't get proper syntax to do that. So, please help me to do so.
How we can set environment variables in core PHP without using any library in Cpanel
6 The reason is that /var/lib/mysql is listed as a VOLUME in the Dockerfile. The changes you make are retained between docker stop <yourcontainer> and docker start <yourcontainer> commands. But when you commit a container, each directory marked as VOLUME in the Dockerfile, ...
I installed latest Docker CS, got a LAMP image from Docker Hub. I'm trying to create a DB in it and make a new image with that DB saved in it. Start the container: docker run --name mycontainer fauria/lamp This starts the Ubuntu-based container and starts Apache server. MySQL server is also running in the container. ...
Docker commit - DB changes are no saved
<div class="s-prose js-post-body" itemprop="text"> <p>Yes. In your compose file, you can specify <a href="https://docs.docker.com/compose/compose-file/#aliases" rel="noreferrer">network aliases</a>.</p> <pre class="lang-yaml prettyprint-override"><code>services: db: networks: default: aliases: ...
<div class="s-prose js-post-body" itemprop="text"> <p>Suppose I have a docker compose file with two containers. Both reference each other in their /etc/hosts file. Container A has a reference for container B and vice versa. And all of this happens automatically. Now I want to add one or more hostnames to B in A's hosts...
How can I add hostnames to a container on the same docker network?
linking is not allowed in AWSVPC. You can do linking only in network mode when its set to bridge. links Type: string array Required: no The link parameter allows containers to communicate with each other without the need for port mappings. Only supported if the network mode of a task definition is set to bridg...
I am new to ECS and I am trying to deploy a couple of containers in a ECS task using Fargate. I have 1 container running that uses Angular2 and is running on nginx, the other container is the backend and is running on Springboot and uses the port 42048. I am using the awsvpc network with Fargate and I have to do it th...
How do I link 2 containers running in a AWS ECS task
No, this is out of scope of the Prometheus project. Look into something like Thanos or Cortex. Both of these solutions provide scalability and high availability. Something like VictoriaMetrics may also be an option if your environment isn't Kubernetes based and you prefer something "simpler" at least regarding the setu...
Can two Prometheus server write to same remote data storage endpoint without any duplicate data created ? I am trying to build setup with 2 Prometheus server for high availability.
Two Prometheus with same data repository
The 'File' class from Java doesn't understand that S3 exists. Here's an example of reading a file from the AWS documentation: AmazonS3 s3Client = new AmazonS3Client(new ProfileCredentialsProvider()); S3Object object = s3Client.getObject(new GetObjectRequest(bucketName, key)); InputStream objectData = object.ge...
I tried to read a file from AWS S3 to my java code: File file = new File("s3n://mybucket/myfile.txt"); FileInputStream fileInput = new FileInputStream(file); Then I got an error: java.io.FileNotFoundException: s3n:/mybucket/myfile.txt (No such file or directory) at java.io.FileInputStream.open(Native Method) ...
How can I read an AWS S3 File with Java?
The answer to this question was:ci/pdf/jenkins-changes.sh LOG="$(cat ".log")" export LOGThen pass it as an argument, instead of a variable:pdf: image: thisimage/this build: context: ../ dockerfile: ./docker/Dockerfile.name args: git_branch: ${GIT_BRANCH} env_file: - variables.env ...
I am running a script in the CI/CD of the pipeline. The goal is to get a string to work with.When I get that result, I save it into a variable and save result in the yaml file of the dockerfile.I am wanting to pass that variable from the CI environment, into the docker-compose container. So, I am trying to export this ...
Send variables into docker container to use in a script
2 Cloning is perfectly acceptable. You don't have to backup to tape... It can be done to a NAS for example, and with the proper security and setup, backups cannot be deleted by unauthorized people. Share Improve this answer Follow ...
I am application developer and don't know much about virtual machine(VM). however, our application is resided on a VM. frequent patch need be apply to fix/update this application. For diaster recovery, It was suggest to backup every thing on the server. so, once server is restored, no application need be re-installed...
Can clone VM be application backup plan?
Try this: docker run -v ${PWD}/script.sh:/home/root/script.sh -w /home/root amazonlinux ./script.sh
I am using the amazonlinux image to test a server setup script without spending time and money on real EC2 instances. However, I cannot get the script to execute within the container. I am running with the following command: `docker run -v script.sh:/tmp/script.sh -w /tmp amazonlinux ./script.sh Which fails with "Erro...
Bash script within Docker container unexpected "permission denied"
0 You can use the -p option to specify the prefix path: nginx -p /home/webapps/confs/ -c nginx.conf Share Improve this answer Follow answered Mar 26, 2020 at 20:12 builder-7000builder-7000...
Trying to set config file. Ubuntu 14.04.3 LTS. User root. nginx -s quit nginx -c /home/webapps/confs/nginx.conf nginx -V ...--conf-path=/etc/nginx/nginx.conf... nginx -s quit nginx -c /home/webapps/confs/nginx.conf -t nginx: the configuration file /home/webapps/confs/nginx.conf syntax is ok nginx: configurati...
nginx -c, set configuration file
1 You are using incorrect project structure. Your project structure is a structure inside a module. You need to push the whole library project to github. Your project should be similiar with this: gradle/wrapper library --- src --- .gitignore --- build.gradle --- prog...
All I have Created My Own Library and I Push That Code in Github, I have Created Library Using JitPack.io but I Face Some Problem When I Call That in My Project Erro is Below : Error:Could not resolve all files for configuration ':app:debugAndroidTestRuntimeClasspath'. Could not find NewInstllSDK.jar (com.github.sa...
i have issue when call my library in gradle