Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
Yes, it's possible. You just have to use theExcludesetting in your_config.ymlto tell Jekyll which files and/or folders it should ignore. Similarly, you can also use theIncludesetting to specifically include files and/or folders.Configuration Settings / Global Configurationhttps://jekyllrb.com/docs/configuration/#global...
Is it possible to have Jekyll on a single subsection of my website hosted on GitHub? i.e. is it possible to have "example.github.io" be a regular site and "example.github.io/blog" be statically generated by Jekyll?I believe it's possible to do this by creating another GitHub repo called "blog" and having that repo mana...
Jekyll static site generation on a subsite?
Using30/15in the minute area should work just fine:30/15 9-18 * * 1-5This will instruct crontab to run your script: “At every 15th minute from 30 through 59 past every hour from 9 through 18 on every day-of-week from Monday through Friday.”ShareFollowansweredSep 20, 2016 at 14:48Alexej MaguraAlexej Magura4,98144 gold b...
I got this crontab code to run a script on crontab every 15 minutes from 9 to 18 all days except weekends. However I would like to start fro 9:30 instead of 9. Is it there a way to do it?*/15 9-18 * * 1-5
Crontab start as half past hour
As suggested by @Mathias in the comments, I updated the script to includeImport-Module -name AWSPowerShelland it worked like a charm.
I am usingNew-EC2Tagto create/update the EC2 tags through Azure Pipeline. I am using "AWS Tools for Windows PowerShell Script" task and below is the code:$Tag = New-Object Amazon.EC2.Model.Tag $Tag.Key = "DesiredInstanceState" $Tag.Value = "Stopped" New-EC2Tag -Resource $instanceName -Tag $TagWhen the task runs I ...
New-Object : Cannot find type [Amazon.EC2.Model.Tag]
This is not possible, SonarQube does accept coverage data from multiple sources, and it thenmergesthem.You should use your individual test/coverage tools to further distinguish between different types of coverage.
We would like to get unit tests and integration tests reported separately in sonarqube 6.+. currently, it shows full tests coverage.
how do we get separate report in sonarqube 6.5+
You can see syslog to find out what happened, seems like invalid command arguments or invalid crontab file format.
My crontab is not working right.. I don't know why, I first stumbled across the time being wrong but now I do date, then set it up 5 minutes ahead and 10 minutes later still nothing happened.. anyone any idea's?My crontab:0,10,20,30,40,50 * * * * /usr/local/bin/flexget -c "/home/user/NASDM/FlexGet/config.yml" --cron 01...
Crontab not working like it should
Absolutely, the aws-sdk is available by default as an NPM dependency inside of the lambda containers so if you leave it as a development dependency your code will still work inside of lambda.Hereyou can see which lambda containers contain which version of the AWS SDK. So in case you really need a specific version or on...
I am quite new in trying to develop lambdas with NodeJs, so this question might sound silly.One of the limitations of lambdas is the size of the function / dependencies (250 MB) and I was wondering ifaws-sdk(which has >45 MB)can be treated as a dev-dependency since it occupies 1/5 of the total size of a lambda.I unders...
Can aws-sdk be a development dependency when developing lambdas with NodeJS?
Asking the OS for memory or returning it back are (relatively) expensive operation because they require a context switch user/kernel and back. For that reason, in most implementations, the malloc call only asks for large chunks and internally allocates from those chunks, and manages freed memory with a free blocks lis...
I'm writing a simple malloc implementation for a college project. One of the tasks is to sometimes give back freed memory to the OS (the example given was of a process using say 1GB malloc-ed memory during a period, and afterwards it only uses 100MB memory until it terminates), however I'm not sure how to implement th...
When should a free() implementation give memory back to the OS?
1 I had the same problem. The issue was in my mistake in parameter "root" in nginx's [your-name].conf file. It should match the path shown when you log in the laradock "workspace" (with "docker-compose exec workspace bash"), e.g. "/var/www". In your case it looks like shoul...
I'm trying to set up Laradock for my Laravel projects. I started with a new laravel project just to test things before trying with my current projects. I followed the tutorial but somewhere during the steps I must have messed something up I think because when I surf to the URL I get a "404 not found nginx". Someone ex...
Laradock 404 not found nginx
# sudo firewall-cmd --add-forward port=port=80:proto=tcp:toport=10080:toaddr=127.0.0.1This command is adding a forward port for the default zone. That will forward ports for traffic that ingress the default zone, e.g. public. This hits thePREROUTINGchains/hooks for iptables/nftables.Your example$ curl -sS http://127.0....
I have a rootless podman container bound to127.0.0.1:10080, and I would like to forward external traffic to port80to go to that container, which I have been unable to accomplish:[me@certvault ~]$ curl -sS http://127.0.0.1:10080 >/dev/null [me@certvault ~]$ curl -sS http://127.0.0.1:80 >/dev/null curl: (7) Failed to con...
Unable to forward traffic to (rootless) podman container with firewalld
If you really just want the subfolder, without your main project to realize said subfolder is pushed, you can:initialize a new Git repository on GitHubclone it locallyadd the files of that subfolder to your local clone, commit and pushThe trick is, in the new local repository, to designate as a working tree your curren...
In my build script in a project (that is in our private Git) it copies all the samples we have to a working folder. I want to push that folder to a public Git project of ours. I don't need it to beregistered at that location as a sub Git project, I just need to push all the files, including subdirectories, up to a dist...
Push a subdirectory in a Git project to a different Git project
Github does not allow you to fork your own repos, your best bet is using an alt to fork your own repo and pushing whenever
I have researched and all the answers I found say to just clone it and push it up under a different repo. Although this may work for many people it does not work for me.I have a framework I made in a repo. I make websites based off of that. I update the framework regularly and need to be able to merge those changes int...
How to fork your own repo in github?
Azure Notification Hub doesn't set this header default. Based on the WNS documentation, by default caching is enabled.As per WNS documentation, when the device is offline, by default WNS will store up to five tile notifications (if queuing is enabled; otherwise, one tile notification) and one badge notification for ea...
I have wanted to implement a way of ensuring that user receive notifications, e.g.Azure NotificationHub - Detect failed notificationsHowever this had limited success when there was many subscribes to a tag, and then ended up with a lot server time.Instead I found this linkcaching Notifications. It specifically states t...
Notification Hub Caching of Toast Notifications
You'll need to update your branch with new commits from master, resolve those conflicts and push the updated/resolved branch to GitHub. git checkout master git pull git checkout <branch> git merge master [ ... resolve any conflicts ... ] git add [files that were conflicted] git commit git push
I merged a PR with the GitHub web interface, which resolved in a successful merge commit. Afterwards, GitHub is complaining This branch has conflicts that must be resolved. Does anybody know why or how to get the status merged?
"This branch has conflicts that must be resolved" but it's already merged
) The problem was in lineerr = kernel.setArg(11, Otr); checkErr(err, "Kernel::setArg()11");I was passing pointer to float Otr instead of passing the corresponding OpenCL buffer Otr_b to setArg function. OpenCL had been trying to copy the pointer to kernel which was expecting buffer. No doubt there was access violatio...
I try to launch at least empty kernel with a lot of parameters but I got memory access violation in function enqueueNDRangeKernel. Please check my code, maybe I missed something or just didn't know.Host code:https://dl.dropboxusercontent.com/u/71256314/main.cppAt first goes some initialization, then I create buffers, t...
Access violation error in OpenCL enqueueNDRangeKernel
Sonarqube doesn't run code coverage instead it imports created coverage report so that you have 2 optionCreate coverage report as is and exclude specific files/folders/methods for Sonarqube.This is calledNarrowing the Focusin official documentation of Sonarqube. You can either exclude setter-getter classes/files or you...
I want to remove the trivial code from the coverage reports for the front end part of an angular project, this means to get rid of the getters and setters. This is how i extract my reports - npm run test-sonar -- --coverage. And everything is included into the report when i open it. Also imports are included into code ...
Is it possible to remove getters and setters from code coverage report in angular
You should run a git pull origin main and merge with your local. When done merging do git push origin main to sync up the remote If your local changes are too large you can use git stash to hide them temporarily and pop them out of the stash (git stash pop). EDIT: October 2020: Github changed master to main. For olde...
I have a master project stored on my development machine. I push this to Github when changes are made. But, I made 2 changes directly on Github as they were only modifications to the README and I was on a different computer. My local is now different to the remote (the Github repo). How do I get the local back in sync...
Github repo out of sync with local
0 Unlike loading data into Redshift, Unloading actually has to run a select statement. Therefore it can't tell us a status like it does when it's loading. e.g if the select statement has to join multiple tables and scan a lot of tables to generate the output then it might...
When you load data to your Amazon Redshift tables, you can check the load status using the table STV_LOAD_STATE. I would like to know if there's a way to achieve the same, but with the unload operation. In other words, I'd like to know if there's a way to find out the current stage of an unload process.
Unload status in Redshift
The targets are defined in thePrometheus configuration. So, as long as you don't change this configuration, the targets will remain there.However, you can use aPushgatewayto collect metrics from ephemeral processes like your Spark jobs. In this case, your Spark jobs would push their metrics to the Pushgateway while the...
I have several spark jobs exposed metrics to prometheus.After the jobs have been finished,the targets in prometheus switch into 'down' state.So will the targets exist forever?If not,can I config prometheus how long it is going to delete the down targets?enter image description here
[Prometheus]What will happen on the targets in prometheus after they are down and never come back
I found an answer to my question. It is a bug in RStudio, albeit not a very well documented one. When there is a space in the .R file name, it is not possible to push the file through git. Removing the space solves the problem. See here for an official comment on the Rstudio board dating February 2015. I am using the ...
I am new to using git and currently employ the following combination: Git-Gui v.0.20 R-studio 0.99.489 R 3.22 Bitbucket code repository Windows 7 Until recently this combination worked fine for pulling and pushing code to the repository. But now I cannot push modified code anymore nor commit it in R-studio. In R...
Using Git in R-studio: cannot stage modified code files
dotnet/core/sdkimage hasapt(notapt-get):$ docker run -ti --rm mcr.microsoft.com/dotnet/core/sdk:latest sh # apt updateFollowingSonarCube documentation, you can use their docker image with theCLIalready installed:image: name: sonarsource/sonar-scanner-cli:latest variables: SONAR_TOKEN: "your-sonarqube-token" SONAR...
I have installed SonarQube on a ubuntu machine via a docker image. All working fine and I'm able to log in without issues.Have connected to our GitLab installation and see all available projects, when I try to configure the existing pipeline with the following, I got stuck.I have the following pipeline.yml in use (part...
SonarScanner fails with apt-get not found
You can't access them because it isn't really overloading, it's replacement. When you define your own::operator new, the old one goes away. That's pretty much that.Essentially, you need to callmallocfrom a custom::operator new. Not only that, but also follow the directions in 18.4.1.1/4 to properly handle errors:Defaul...
Suppose I need to overload global::operator new()for storing extra data with each allocated object. So basically it would work this way:for each call to global::operator new()it will take the object size passed and add the size of extra datait willallocate a memory blockof size deduced at previous stepit will offset th...
How do I call the original "operator new" if I have overloaded it?
This error comes when there is a heavy load on the server. First I had tried by increasing the value of worker_connections but it didn't work. Queue size for uWSGI is by default 100, so when more than 100 requests from Nginx to uWSGI is passed, queue get full and Nginx throws 502 to the client, to solve this increase t...
I've django app host using nignx-uwsgi. Here is my uwsgi configuration:[uwsgi] master = true socket = /var/uwsgi/uwsgi.sock chmod-socket = 666 chdir = /home/ubuntu/test wsgi-file = /home/ubuntu/test/test/wsgi.py virtualenv = /home/ubuntu/virtual vacuum = true ...
Resource temporarily unavailable using uwsgi + nginx
Two things can help you here. If your friend's changes are on a branch, and you want to merge the branch to keep his changes alive in the graph, you can git merge -s ours his-branch If his commit is on master, already pushed, and just needs to go away, you can git revert his-commit ...making a new commit that un-do...
I work collaboratively with a friend on a personal project, and he maintains a separate GitHub repository from me. I am marked as a contributor to his repository. I have been making some changes ahead of him, and they render one of his commits unnecessary - we don't need the code. I have already committed my changes t...
How can I merge with a remote repository, and undo what it adds?
<div class="s-prose js-post-body" itemprop="text"> <p>Since <a href="https://github.com/moby/moby/blob/master/CHANGELOG.md#1130-2017-01-18" rel="noreferrer">Docker 1.13.x</a> you can use <a href="https://docs.docker.com/engine/reference/commandline/container_prune/" rel="noreferrer">Docker container prune</a>:</p> <pre...
<div class="s-prose js-post-body" itemprop="text"> <p>This question is related to <em><a href="https://stackoverflow.com/questions/17014263/should-i-be-concerned-about-excess-non-running-docker-containers">Should I be concerned about excess, non-running, Docker containers?</a></em>.</p> <p>I'm wondering how to remove o...
How to remove old Docker containers
Mod_rewrite can't really internally rewrite URLs across domains, though itcouldproxy them (usingPoption in RewriteRule). Assuming that the domain is the same, you could do something to redirect the client's browser to a friendly URL if the old one is used while internally rewriting the friendly URL back to the old one,...
I think this is a very stupid question so I apologise, as i think i may completely misunderstand mod_rewrite.Say you have a URLwww.domain.com/products/item.php?id=1234mod_rewrite can rewrite that to a friendly URLwwww.domain.com/products/item/1234(for example)So, if i type in wwww.domain.com/products/item/1234 this wil...
Does mod_rewrite only translate external requests to internal files and not vice versa?
The Apache QSA flag is query-string-append, in the Nginx world the query string is stored in $args. So with try_files something like that should work: location ~ ^/(?<page>[^\/]+)$ { try_files $uri /index.php?page=$page&$args; } Now I do not understand all the things you do in the location 'regex', (I'm a weak hum...
I want to convert some apache rules into nginx ones, and im new to nginx ... So I searched and searched; but didn't find what I want to do. Here is the lines I want to convert in Nginx : RewriteCond %{REQUEST_FILENAME} -s [OR] RewriteCond %{REQUEST_FILENAME} -f [OR] RewriteCond %{REQUEST_FILENAME} -l [OR] RewriteCond...
How to do like apache QSA flag with nginx when you use try_files?
This allocates an array of n pointers to int. A pointer to the first element in this array of pointers is stored in v. It is a double pointer, such that accessing an element via v[i] returns a stored pointer from the array.
int **v = new int*[n]; I'm confused as to what this does? could someone please explain?
How is memory allocated in this line of code "int **v = new int*[n]; "?
Check out the apache config files. For Debian/Ubuntu theyre in/etc/apache2/sites-available/for RedHat/CentOS/etc they're in/etc/httpd/conf.d/. If you've just installed it, the file in there is probably nameddefault.Make sure that the config file in there is pointing to the correct folder and then make sure your scripts...
I've seen this question answered many times, but most end either unanswered or by telling the asker to put this:<?php phpinfo() ?>in a test file. Obviously, if that produced what was expected, I wouldn't be here. Instead, I get a 404 error.I'm using an ubuntu 12.04 server with Amazon. Apache is installed, php5 is insta...
Why is php not running?
You need to change the permission of the bash file bychmod +x entrypoint.shbefore calling ENTRYPOINT. So change your code to the following:USER airflow WORKDIR ${AIRFLOW_HOME} RUN chmod +x entrypoint.sh ENTRYPOINT ["/entrypoint.sh"]Rebuild the image and run the container, it should work.
I am trying using Docker using Dockerfile.My Dockerfile as follows, where I am using debian linux system.FROM debian:jessie ENV DEBIAN_FRONTEND noninteractive ARG AIRFLOW_VERSION=1.7.1.3 ENV AIRFLOW_HOME /usr/local/airflow .. .. COPY script/entrypoint.sh /entrypoint.sh COPY config/airflow.cfg ${AIRFLOW_HOME}/airflo...
getting permission denied in docker run
I think only way is to create a new bucket with correct name and then copy all your objects from old bucket to new bucket. You can do it using Aws CLI.
After all the tough work of migration etc, I just realised that I need to serve the content using CNAME (e.g media.abc.com). The bucket name needs to start with media.abc.com/S3/amazon.com to ensure it works perfectly. I just realised that S3 doesn't allow direct rename from the console. Is there any way to work aroun...
How to rename AWS S3 Bucket
The original NSString will be released by the system and so, won't cause any leak.
I am trying to wrap my head around NSString and NSMutableString, and how they affect memory. In my research, I've concluded that if I create an NSString object and give it a value, then change the value later, the original object is replaced by another with the new value. My question is, in the case of changing the va...
NSString and NSMutableString memory considerations
If your Jenkins server doesn't have access to the internet to download the update file, then it's not going to be able to auto-install the scanner either.For your case, you'll need to install a scanner manually, then configure the path to it in Jenkins.
I am following the instructions at the url below to configure jenkins/sonar.https://docs.sonarqube.org/display/SONARQUBE52/Installing+and+Configuring+SonarQube+Scanner+for+Jenkinsand I come to that stepScroll down to the SonarQube Runner configuration section and click on Add SonarQube Runner. [...]If you don't see a d...
Installing and Configuring SonarQube Scanner for Jenkins - needs clarifications
First, no need to use thegit://protocol: http(s) works just fine;git clone https://github.com/Varying-Vagrant-Vagrants/VVV.git vagrant-localSecond, you will be creating vagrant-local in the current folder you are in, so make sure to do first:cd %USERPROFILE%Ne need for a git bash here, but should you be in a git bash, ...
I'm trying to install varying-vagrant-vagrants (vvv).When I enter the git command:$ git clone git://github.com/Varying-Vagrant-Vagrants/VVV.git vagrant-localI getfatal: could not create work tree dir 'vagrant-local': Permission deniedI tried going in the security settings for my user folder and setting permissions for ...
Windows 10 file permission error with Github
You dont want to run the CRON every minute instead create a specific job for your user and run it once. Something like this should work (untested - on mobile)$your_users_date; $cmd = "sudo crontab -l | { cat; echo ". date("i H d m",strtotime($your_users_date))." * php /path_to_your_script.php arg1 arg2; } | crontab -";...
I want to create a cron job which will run with the user define time in database. For eg.User can set start time and end time in the database .When the end time is reached I want cron to trigger one script to send mail.First of all Is it possible in cron or I have to go with some different approach ? and this all thin...
Create a cron job with user define dtime
The containers in same pod shares the localhost, so you need not link containers, just use localhost:containerPort.
I am trying to see if there are any example to create a Kubernetes POD which starts 2-3 containers and these containers are linked with each other but couldn't find any.Does anybody tried linking containers using Kubernetes config.
Kubernetes configuration to link containers
24 Running "bundle update" worked for me. Share Improve this answer Follow answered Feb 9, 2017 at 6:11 WaqasWaqas 97911 gold badge88 silver badges1717 bronze badges Add a co...
An error occurred while installing json (1.8.1), and Bundler cannot continue. Make sure that gem install json -v '1.8.1' succeeds before bundling. xcrun: error: invalid active developer path (/Library/Developer/CommandLineTools), missing xcrun at: /Library/Developer/CommandLineTools/usr/bin/xcrun make failed, exit cod...
An error occurred while installing json (1.8.1), and Bundler cannot continue - Rails
It would be:0 55 23 1/1 * ? *There is a nice website exactly for your case:CronMakerCronMaker is a utility which helps you to build cron expressions. CronMaker uses Quartz open source scheduler. Generated expressions are based on Quartz cron format.
I am using Quartz for Scheduling my job in java. I have used "CronTrigger" for setting my time. I want to fire my Job each day at 11:55 Pm in night. What should i write in the setCronExpression(" ") for having my Job Done. .??What i thought of the Code is:---CronTrigger trigger = new CronTrigger(); trigger.setName("run...
Java Scheduler Quartz Cron Trigger Time Setting
3 Stream the input to a file. Really, there is no other choice. It comes in faster than you can process it. You could create one file per second of input data. That way you can directly start processing old files while new files are being streamed on the disk. Share...
I am working on a project where we can have input data stream with 100 Mbps. My program can be used overnight for capturing these data and thus will generate huge data file. My program logic which interpret these data is complex and can process only 1 Mb data per second. We also dump the bytes to some log file after ...
How to handle 100 Mbps input stream when my program can process data only at 1 Mbps rate
0 You can print it and see if you want. It would be something like this: LambdaContext([aws_request_id=7d78a745-bb86-42b8-89f0-389e9bdc3dcfe,log_group_name=/aws/lambda/mldx-devops-v1-TimeLambda-hINHLpDGpaNR,log_stream_name=2022/09/01/[$LATEST]0553dabc81194a57adf2a96775f88...
I set sample lambda function as follows to detect what is context. This function is hooked by API gateway. import json def lambda_handler(event, context): return { 'isBase64Encoded': False, 'statusCode': 200, 'headers': {}, 'body': json.dumps(context) } And when I send GET req...
How to see context variable in lambda function
1 Here's some instructions, but before you do anything, MAKE SURE YOU HAVE BACKUPS OF EVERYTHING. To delete a remote branch at origin (assuming origin is github): git push origin :my-branch You should also be able to delete master that way. Then you can create a new branch...
I encountered with troubles in github. When I enter my valid credentials it asks me to enter the code I received on my email. The problem is that I have forgotten my email password and cannot recover it. I still can push commits. Now I would like to make the repositories I have empty. How can I do it?
How to empty the repository using push github
git remote rm origin This command will delete the local reference to the remote repository. Intellij should be smart enough to pick up on that and stop fetching. If that doesn't work, try going to the menu option listed here: How to change GitHub repository in IDEA Intellij?
I deleted some repository from GitHub. and IntelliJ IDEA keeps trying to Auto fetch them. I'm constantly I get this error message: 9:06 PM Auto fetch failed origin: remote: Repository not found. Authentication failed for 'https://github.com/Guilderm/skypos.git/' Development: remote: Repository not found. Authentica...
IntelliJ IDEA keeps trying to Auto fetch a deleted repository
It very depends what you are planning to do further in the code. Setting a pointer to to NULL/nullptr after deleting the object it used to point to (regardless whether it is a pointer pointing to the array or to the singular variable), is done purely to mark a pointer as deallocated ,so you can easily check if that po...
I'm studying about new and new[] from learncpp and there is one small detail in the site, which isn't clear enough for me. The following picture will explain my question: So do we have to set the array-pointer also to 0 after we have deallocated it like in the single variable case? Is this a requirement or is the p...
Difference between deallocating a single variable or array in C++
RewriteRule ^profile/([0-9]+)/([A-Za-z0-9-]+)/?$ index.php?p=profile&id=$1Should work for :www.domain.com/index.php?p=profile&id=20 to www.domain.com/profile/20/profile-friendly-name
I am trying to rewrite the urls of a site, i should mention that the way index.php works now is getting the p (page) parameter and including the appropriate file.So requesting a page is like this:www.domain.com/index.php?p=home www.domain.com/index.php?p=search www.domain.com/index.php?p=profile www.domain.com/index.ph...
Creating SEO friendly urls using htaccess
Here's a tool that continually checks and potentially merges GitHub pull requests. Developers can communicate with the tool through comments on the pull requests.https://github.com/xen-org/pull-request-manager
Is there any way to set up github web hooks in the way merge happens only in the case autotests passed on merged version?I.eI receive pull requiest and review ituse 'merge pull requiest' buttonmerge details are sent to my serverlocal merged version created locallyautotests runautotests result is sent back to githubmerg...
Run autotests as a result of clicking 'merge pull request' button
5 I ran into this issue as well, CognitoUser was missing the attributes object. The solution was to include the aws.cognito.signin.user.admin scope both on the client-side configuration and on the App Client Settings page in the User Pool settings. The oauth configuration...
I have an Angular 11 project in which I'm using AWS Amplify (aws-amplify v3.3.26) and I'm using Cognito user pools to manage users. I've set up the hosted UI and I've not used any custom attributes. Everything is working and I've selected all the attributes for read and write permissions in the user attributes in the ...
Unable to get user attributes in the cognitoUser object after authenticating using aws-amplify
It's not possible.Thedocumentationstates:"If the border mode is specified instead, texture fetches with out-of-range texture coordinates return zero."
When one is looking up texture values in cudaAddressModeBorder mode the values outside the texture range are set to 0. Is it possible to specify a custom value for the out-of-range texels? For example, for the application I'm working on the value 1.0 would make more sense.
Is it possible to customize the out-of-range texture value in cudaAddressModeBorder mode?
You need to make p12 (pfx) file from .key and .crt. You can use openssl do make it.openssl pkcs12 -export -out certificate.pfx -inkey privateKey.key -in certificate.crt -name "friendlyname_for_certificate"Then import certificate.pfx into LocalMachine\My store. It should be available in IIS.
I have .crt file and .key file and I want to use SSL for my web application. I have tried with option "Complete Certificate Request" to import certificate in IIS, but it show certificate imported and actually it is, but when I edit binding, it won't display in selection in IIS.My certificate does not have private key, ...
Using SSL with IIS
I was also facing same issue the solution is very simple, bind your inspector to 0.0.0.0 instead of 127.0.0.1so change your package.json script to something like this:scripts:{ "debug": node --inspect=0.0.0.0:9229 ./bin/www ... }Reference:https://medium.com/@auchenberg/introducing-remote-debugging-of-node-js-apps-on-...
I have a node.js application hosted in AWS EC2 instance. I ssh into the host with .pem file for authentication. Is there any way I could debug this code in VS Code, I see there isremote debuggingin VS Code, there is configuration to specify port and host but no option to specify the pem file.How should I configure VS c...
Remote debugging in Visual Studio Code
7 This SO question provides some answers: git log --author="<authorname>" --pretty=tformat: --numstat and for the graph itself, I know about this jQuery program (but I haven't tested it), based on json-formatted data. Illustrated here, reported by this message) Sha...
Is there some program out there (or some magical git plugin I've missed) to get an impact graph or something similar from a git repo without going through github? I'm unsure where I'd begin exactly to write my own, as far as data collection goes (I can generate graphs). I assume there's some flags I could pass to git-...
How to generate a github-like Impact Graph?
Watch this: http://www.youtube.com/watch?v=_CruQY55HOk Then install MAT for Eclipse. Happy hounting!
Closed. This question is seeking recommendations for software libraries, tutorials, tools, books, or other off-site resources. It does not meet Stack Overflow guidelines. It is not currently accepting answers. We don’t allow questions see...
Android: tools to find memory leaks? [closed]
Great question. The answer isNO, they do not share the sameAutoPurgingImageCache.If you take a look at theImageDownloaderinitializer, you'll see that it creates a customAutoPurgingImageCacheautomatically.public init( configuration: NSURLSessionConfiguration = ImageDownloader.defaultURLSessionConfiguration(), do...
I use AlamofireImage in my project a lot like this to set an image to an UIImageView:let imageView = UIImageView(frame: frame) let URL = NSURL(string: "https://httpbin.org/image/png")! imageView.af_setImageWithURL(URL)At some point in my app I need to get fetch an image directly, because I want to do something else th...
Alamofire: Do af_setImageWithURL and ImageDownloader share the same AutoPurgingImageCache?
Cloud Run is a serverless product. As serverless product, you don't manage the servers and the network. Therefore, asking for the podIP make no sense and has no value (at least, it will be a private IP in the Google Cloud serverless network).So, the answer is "NO, you can't get the podIP".But why do you need it? You ca...
I am trying to run a container image in GCP Cloud run.Tech Stack: .NET 6, C#, Orleans.NET, ASP.NETMy code requires the podIP to work properly.This programs works properly in a regular Kubernetes cluster, because I am able to set an environment variable with the assigned podIP with the following configuration in the "en...
How do I get the status.podIP within the container program in GCP Cloud Run?
I'm writing this answer for any future people which come by this. If there is no NPM error message. The issue is memory exhaustion there are two ways to resolve this on Beanstalk specifically: Use a machine with more memory (instead of say a t2.micro) (recommended) Switch to yarn. Yarn uses much less memory than npm....
I am using elastic beanstalk from AWS to deploy and host a react application using Docker. The deployment fails mysteriously, even though I succeed building the Docker image on my local machine. An inspection to the logs on the EC2 instance (using eb ssh) does not provide more explanation that this: -----------------...
Fail to deploy node on Elastic Beanstalk : The command '/bin/sh -c npm install' returned a non-zero code: 1
To get to the compare view, append /compare to your repository's path. Every repository's Compare view contains two drop down menus: base and compareTo compare the commits in timeframe you can do type the branch name in the compare dropdown, followed by a@, and then the date wrapped between a{ }notation.Here are two ex...
How can I know the total number of commits done on a project through the GitHub web interface?How can I get the total number of commits in a specific time frame?What I can get now is the number of commits done by each developer, not the total one.Example:https://github.com/BVLC/caffe
Total number of commits on git/GitHub in a specific timeframe?
You must push your changes to the repository.Assuming "develop" is the branch's name (on the upstream repository)Usegit push origin develop
I followed this hallowedpageto sync my fork but the fork is still not completely synced. I am working on "develop" branch and the original repo had newly added "develop_backup" branch. But after syncing the fork still the newly added branch also is not reflected; neither are the latest changes.These are the steps I did...
Fork is not completely synced even after syncing
37 Try this git command git push origin master --force Share Follow edited Feb 20, 2018 at 7:19 answered Feb 16, 2018 at 11:35 harish samiharish sami 48544 silver badges99 bron...
I am new to Git and I am trying to merge my branch (With_Base_Class) with the remote master branch from within Xcode. I tried commiting and pulling in every possible order (I hope), but every time I recieve the following error: The local repository is out of date. Make sure all changes have been pulled from the remot...
The local repository is out of date. Make sure all changes have been pulled from the remote repository and try again
Your script is running: echo "$DOCKER_REGISTRY_PASSWORD" | docker login -u "$DOCKER_REGISTRY_USERNAME" --password-stdin Which looks correct to me. Make sure that: You have added a repository secret with the docker registry password (for consistency's sake, let's use DOCKER_REGISTRY_PASSWORD) You provide that secr...
I have a project where I'm having a shell script that pushes images into the Docker hub after a successful build. In the script, I have the following defined: echo "Attempting log in to $DOCKER_REGISTRY_URL" echo "$DOCKER_REGISTRY_PASSWORD" | docker login -u "$DOCKER_REGISTRY_USERNAME" --password-stdin When the p...
GitHub Action Fails for Docker Registry Login
9 Try below solution which works for me: Go to Windows Credential Manager. This is done in a EN-US Windows by pressing the Windows Key and typing 'credential'. In other localized Windows variants you need to use the localized term (See comments for some examples). alterna...
I just wanted to make a clone of a private repositoy from Github, and the prompt showed up and asked for my login details. I made a typo, but I don't get a second try. If I try to clone again I get the following message: Cloning into 'name'... remote: Invalid username or password. fatal: Authentication failed for 'ht...
GIT entered wrong password, can't clone anymore
It is very common to run Emacs locally (e.g. on your Mac) and edit files on remote systems usingTRAMP, an excellent built-in library.To edit a remote file over SSH,find-fileusing a pattern like//ssh:user@host:path/to/fileIn this casepath/to/fileis a path on the remote system relative to your home directory. As you migh...
I am new to emacs and was trying to use it when I am editing files on an AWS server. The problem is that when I ssh from terminal (on my Mac) and try to use the Meta or Esc keys they don't work. The meta key just causes characters like this --> √≈ß to appear. The esc key causes nothing to happen. Does anyone know how ...
Using Emacs on AWS Ubuntu system - Meta and Esc keys don't work
It depends on how you defined the cron job, from the dochttps://developers.google.com/appengine/docs/python/config/cron#Python_app_yaml_The_schedule_format:By default, an interval schedule starts the next interval after the last job has completed. If a from...to clause is specified, however, the jobs are scheduled at r...
I am wondering whether Cron Jobs ran by Google App Engine wait for the last Cron Job to finish before starting a new Job? What is happening if:A Job is schedules to run every minute but one Job instance takes longer than a minute to complete?Two different Jobs are schedules to run at the same time (say, every minute)?W...
Will GAE Cron Job wait for previous Cron Job to be completed before running?
GitKraken can do this for you out of the box by just cloning the forked repo and selecting GitHub as the issue tracker for the selected repo. All issues will then appear on the sidebar.
This probably sounds weird but I would like to fork someone's repo with all the issues attached to it. I know I can fetch changes from the original(upstream) repo such as added files or changed files but I would like to sync issues from the original repo with my repo. I tried to sync my fork with the upstream but I ...
Is it possible to fork github repo with original issues?
What are the results of nslookup google.com in the container and the node? If the pod's dnsPolicy is ClusterFirst, google.com DNS query should be forwarded to the upstream DNS which the node specified. It would be also useful to show the kube-dns container config and logs.
I installed kubernetes by following this tutorial. One of my containers tries to get resources from an external domain, such as google.com. But it fails because kubernetes dns doesn't use external name resolving. How can I configure kubernetes using dns 8.8.8.8 ?
Kubernetes dns external resources
This works on my scenario, but I hope there will be a better solution: I set the network aliases of the backend to the domains I need to use so that it goes through the proxy: networks: default: aliases: - auth-xx-dev.pixel.local - listing-xx-dev.pixel.local Reference: docker reverse...
I am new to docker, so I'm not really sure if I am correct on everything, but I will try my best to describe my situation. I have two separate docker containers, one for the API and one for the frontend project. I used jwilder/nginx-proxy so I can access them both using hostnames(VIRTUAL_HOST environment variable on d...
Docker container cannot access another container via VIRTUAL_HOST environment variable
You can add a GitHub workflow to your project that will run on every opened pull request. In that workflow you can compare the source and target branch and throw an error if it's not meet your requirements. The workflow can be something like this.name: Check pull request source branch on: pull_request_target: typ...
I am trying to define the following rule: "allow pull requests tomainonly fromdev"; so if someone opens a PR from a random branch tomainit should not be allowed. How do I go on doing that?
In github is there a way to prevent pull requests to main from any other branch but dev?
You just need thissingle rulein theDOCUMENT_ROOT/.htaccessfile of olddomain:RewriteEngine On RewriteCond %{HTTP_HOST} ^(www\.)?olddomain\.com$ [NC] RewriteRule ^ http://%1newdomain.com%{REQUEST_URI} [R=301,L,NE]Explanation:NC- ignore caseL- LastR=301- Send301status to browserNE- no escaping%1- is the value we capture ...
Essentially I need an .htaccess file that will redirect all traffic to our new domain. It need to work in the following conditions:http://www.olddomain.com/path/file.php => http://www.newdomain.com/path/file.php https://www.olddomain.com/path/file.php => http://www.newdomain.com/path/file.php(note in the above case the...
.htaccess redirect to new domain with path intact for both www and non-www as well as https and non https
5 To retrieve your inventory, you could do something like this: import boto.glacier c = boto.glacier.connect_to_region('us-east-1') vault = c.get_vault('myvault') job = vault.get_job('<your job id>') response = job.get_output() print response.read() It's a bit confusing...
I'm actually using python boto to store data on my glacier's vault and launch retrieve jov and inventory job. This works quite well. But i didn't find any example about the method to use to download an archive from glacier? I had launch retrieve jov and inventory job and got request id 4 hours later etc, but howto gra...
Boto AWS Glacier - Retrieve archive
Those modules are not available by default. I found this (older) list of available modules, along with code to generate a current report of what is available, should you wish to do so:https://gist.github.com/sjehutch/36493ff674b1b9a16fc44a5fc270760cYou can build your own environment using a virtualenv and upload that t...
For example, I would like to use scikit-learn for machine learning, and pandas, numpy, and matplotlib/seaborn for data analysis. Does AWS Lambda support all these libraries? Is there a list of libraries supported? Or should I create a virtual server instead, to ensure I can use any Python library I want?
Does Amazon Web Services (AWS) Lambda support all Python libraries?
The problem was that PassengerLoggingAgent wasn't installed when I did 'gem install passenger'. I installed it as per this discussion (by grabbing it from the standalone zip) and everything is fine now: http://groups.google.com/group/phusion-passenger/browse_thread/thread/323dc1c7b0e5ee11
On Ubuntu with REE 1.8.7, with nginx 0.8.53 manually compiled with the Passenger 3 module doc installed as per the docs, I get an error in the nginx error log when nginx starts: Unable to start the Phusion Passenger watchdog because it encountered the following error during startup: Unable to start the Phusion Passeng...
Why does Passenger 3.0 with nginx crash on startup(?) then 403s?
If you want to run uWSGI as particular user, there are only 2 options:run uWSGI server directly from this userrun uWSGI as root and add uid and gid options.
I have Django setup in NGINX + uWSGI. I'm able to get it running fine under my current logged in user (with help from aquestionI asked few days back) but now I want torunuwsgi --ini uwsgi.inias a limited-access user.Here is what I've done so far:1. Created a userdjangouserwithout login access and without a home directo...
How do I run uWSGI as a limited-access user?
Thevalues.yamlfile in the root of thestable/airflowrepo has an example of how to do this:airflow: extraConfigmapMounts: [] # - name: extra-metadata # mountPath: /opt/metadata # configMap: airflow-metadata # readOnly: true # # Example of configmap mount with subPath # - name: extra-metadata # m...
According tothe stable/airflow docsyou can:mount a file to /requirements.txt to get additional Python modules for your DAGs to be installed on container start. Use the extraConfigMapMounts configuration option for this.But the README is a bit thin on documentation for this option:| airflow.extraConfigmapMounts | Addit...
Make local requirements.txt file available to microk8s stable/airflow
Although I do not fully undestand the reason behind it, it seems like the --no-cache-dir option was causing the issue. The dockerfile below builds without an issue: FROM debian:buster RUN apt-get update && \ apt-get install -y \ python3 \ python3-pip RUN python3 -m pip install PyInstaller==3.5 RUN pytho...
I am building a docker image. Within it I am trying to install a number of python packages within one RUN. All packages within that command are installed correctly, but PyInstaller is not for some reason, although the build logs make me think that it should have been: Successfully installed PyInstaller The minimal Doc...
No module named PyInstaller' after what appears to be a successful install
/has a higher precedence thanand, so try:( node_memory_MemAvailable and on (instance) node_meta{node_name="dc0-docker6"} ) / ( node_memory_MemTotal and on (instance) node_meta{node_name="dc0-docker6"} )
I am e newbie with prometheus, I'll try to explain my problem clearly.The 2 following queries return data in Promotheusnode_memory_MemAvailable{instance="10.0.0.15:9100",job="node-exporter"} node_memory_MemTotal{instance="10.0.0.15:9100",job="node-exporter"}And then I can divide them:node_memory_MemAvailable{instance=...
Division results in no data when using the 'and" operator
The Django template engine has basically three steps to perform: load the template file from the filesystem compile the template code into python execute the code to output plain text (usually HTML markup). The cached.Loader caches only the two first steps : your templates wont be loaded and compiled every time, but...
ERROR: type should be string, got "\nhttps://docs.djangoproject.com/en/dev/ref/templates/api/#django.template.loaders.cached.Loader\nHere we have \"cached.Loader\" to cache template\nhttps://docs.djangoproject.com/en/dev/topics/cache/#template-fragment-caching\nAnd we also have \"Template fragment caching\"\nI know the latter allows a finer control over which parts to cache. But if I enable both, will it consume double amount of memory for same fragments?\n"
Django : Two ways of caching template : What is the difference?
Skaffold is a tool for deploying to any Kubernetes cluster, agnostic of cloud vendor. Skaffold can be used with any Kubernetes cluster, whether it is hosted on a cloud provider like Google's GKE, Amazon EKS, or running on-premises. Skaffold does not have any specific limitations regarding shared volumes or other resour...
I exploring using Skaffold with our EKS cluster and I wonder if the tool is agnostic to the cloud vendor and if he can work with any k8s cluster?Does he have any limitations regarding e.g shared volumes and other resources?
Does Skaffold has any limitations works with EKS
Assuming the current repository state is this:server: A---B---C---D ^master workstation: A---B---C---D ^HEAD ^master ^origin/masterAnd the desired state is this:server: A---B ...
What is the best way to remove commits from GitHub without affecting my local repositories since i made some commits that i wish to keep?Github repository has two wrong commits that i desire to remove while keeping the commits on my local repository.
GitHub - Remove pushed commits without affecting local repository
2 This is a terrible idea. Please do not do it. Who'd trust an engineer who violated confidentiality just for the green squares? "My current employer uses GitLab" should be plenty of explanation why your GitHub profile doesn't have a lot of green squares. (In this case, it'...
I'm currently employed at a company that uses GitLab for version control and repository management. However, many job applications require a GitHub profile as part of their criteria, and my primary platform for development work is GitLab. Consequently, my GitHub profile appears inactive and devoid of contributions. Is...
How can I mirror my GitLab commits to GitHub for a more active profile?
You can refer to the official GitHub "Error: Repository not found" listing the main reasons:spellingpermissionsssh access (that is, if you have used an ssh url)existence of the repoIn your case, the origin url is wrong:git remote set-url origin https://github.com/USERNAME/REPOSITORY.gitIn your case:git remote set-url o...
I cloned my repo on my laptop and followed the steps:git add . git commit -m "First commit" git remote add origin (repository URL) git push -u origin masterbut for some reason, I get a fatal error telling me that the repo not found. What's the issue?
Push files to Github
You can intercept calls between containers usingPnTestwhich uses mitmproxy under-the-hood but automatically re-configures your containers to route all outbound traffic through a proxy.
I am looking for a way to intercept all the HTTP requests between docker containers, for debugging purposes.I am usingMitmProxyand I can intercept the requests done from the containers to the internet (configured theHTTP_PROXYandHTTPS_PROXYenvironment variables), but not the ones done internally to other containers in ...
how can I interpect all requests between docker containers?
You can useprom2teamswhich is a Python webserver that receives alerts from Prometheus Alertmanager and forwards it to Microsoft Teams.You can install it with pip3:pip3 install prom2teams
I am setting up the monitoring for the Linux server and send alerts to Microsoft teams. I have set up the Prometheus monitoring. Now how do I send alerts to teams? I have installed everything on the centos 7 machines. Can anyone please guide me in this.
Alert manager/prometheus alert to Microsoft teams
Followed the above guide and it worked. https://medium.com/@francisyzy/create-aws-elb-with-self-signed-ssl-cert-cd1c352331f Strangely, I've succeeded in uploading the certificate into IAM using the command above: AWS CLI: aws iam upload-server-certificate --server-certificate-name CERT_NAME --certificate-body file://p...
I've created an ALB using Boto3 and want to configure that load balancer work on HTTPS (self-signed). In order to do that, I have to generate an SSL certificate with open-ssl: openssl req -x509 -nodes -days 365 -newkey rsa:4096 -keyout key.pem -out cert.pem Then, I've imported the certificate into AWS ACM with no pr...
AWS self-signed Application Load Balancer
Youcan't decreaseallocate storage for your RDS. It can only be increased. Fromdocs:You can only increasethe allocated storage. Youcan't reducethe allocated storage.You would have to migrate your database to new RDS instance with smaller storage.
I have a RDS instance in amazon I want to modify the allocated storage for, it's currently at 500GB but I want it at 50GB. However when I try changing this in the console I receive the following error:Invalid storage size for engine name postgres and storage type standard: 60I have a Elasticbeanstalk application that h...
AWS: Invalid storage size for engine name postgres
2 Tensorflow officially only supports CUDA, which is a proprietary NVIDIA technology. There is one unofficial implementation using openCL here which could work, or you could try using Google colab Share Improve this answer Follow ...
I want to use TensorFlow on Windows (Win 10) with a AMD GPU. If I google, there are a lot discussions and sources but I just couldn't figure out what's the best way to do this at the moment. Could someone write a short installation instruction that he thinks is the best and most up-to-date way of doing so?
How can I use TensorFlow on Windows with AMD GPU?
When doingnpm installfor your project you shouldn't have to run it as the superuser.sudois normally use with the-gflag (global) ofnpm install.Looking at the error message, it seems like whichever process that is running the gitsubmoduleprogram is having issue in creating files in the/root/.npm/...directory. It is proba...
[sudo] password for mayukh: npm ERR! code 128 npm ERR! Command failed: /usr/bin/git submodule update -q --init --recursive npm ERR! fatal: failed to stat '/root/.npm/_cacache/tmp/git-clone-02bd6ef9': Permission denied npm ERR!npm ERR! A complete log of this run can be found in: npm ERR! /root/.npm/_logs/2018-12-28...
I wrote this command "sudo npm install --save-dev grafana/grafana-sdk-mocks", I am getting this error
If you look in your project folder you should see a user settings file called<projectname>.csproj.user.Within this file it may look like the following<?xml version="1.0" encoding="utf-8"?> <Project ToolsVersion="15.0" xmlns="http://schemas.microsoft.com/developer/msbuild/2003"> <PropertyGroup> <ActiveDebugProfile...
I've installed Visual Studio 2017 Community RC with .NET Core and Docker (Preview) so I could try the "Add docker support" on my project.Unfortunatelly I couldn't get things work together (win8.1 + docker toolbox + hyperv engine + docker tools seems like not works together well) so I decided to remove docker support fr...
How to remove docker support from an ASP.NET Core project?
At first, avoid usingsudoin containers (use for examplegosuinstead).I'm bringing bad news, your conclusion is correct. I'm facing same problem with/dev/dri/card0which belongs toroot:video. But there is the difference in GIDs. Groupvideohas GID91on host system (Arch Linux) but different GID44in container (Ubuntu:20.10)....
I was trying to setup sound inside a Docker container, so I've created an image with alsa utils installed and a user that belongs to theaudiogroup. As pointed out in docs and multiple tutorials, I ran the image with--device /dev/snd. The result was that I could play sounds without issues, but only when running the comm...
How to ensure proper relationship between host GIDs and docker container GIDs?
Simply typing the exit command will log you out of your instance. Hope this answers your question.
I would have thought this would be a question I could find the answer to with a quick Google search, but I can't. I have connected to my EC2 instance via an SSH client but now I would like to disconnect, but I have no idea how.
How do I disconnect from my EC2 Instance?
1 You can create a submatix header for your buffer using cv::Mat::operator(). Pass a cv::Rect for ROI you want to process in current loop iteration ({0, 0, i, i} in your case) and it will return a view of your buffer region as another cv::Mat instance. It will not allocate...
I am rewriting some legacy code that does matrix operations on doubles using a raw C-style array. Since the code already has a dependency on OpenCV somewhere else, I want to use the cv::Mat class instead. The specific code that bothers me works on square matrixes from size 1*1 to NN. It does so by allocating an NN buf...
How can I reserve a fixed buffer size in OpenCV Mat?
Trycertreqtool. Data are given in a form of inf file. The before mentioned link contains information about the file structure.Thislink contains example inf file of an SSL certificate (could be of more help).Then runcertreq -new your_inf_file.inf generated_req.reqAs a side effect it will generate a self signed certifica...
Is there a way to create a self signed certificate in IIS in windows 2012 R2 using command prompt? I went many documents, everywhere there is tutorial using Windows powershell but not using command prompt. I do not have makecert utility installed in my system. Is there any other command or utility. I saw a selfssl uti...
How to create a self signed cerficate using command prompt?
It's an unexpected xdebug remote host behavior. I have fixed it with stopping using an environment variable. In my docker-compose.yml app: ... environment: ... # removed the line below XDEBUG_CONFIG: remote_host=docker.for.mac.localhost ... And added it to t...
I'm working with Docker through the docker-compose commands. I want to be able to run my app both in a debug and a normal mode, but now the debugger starts anyway. I have such an app.docker file. FROM php:7-fpm RUN apt-get update && apt-get install -y libmcrypt-dev mysql-client \ && docker-php-ext-install mcrypt...
XDebug. Docker. Can't disable an autostart
For cron job lines in a crontab it's not required to specify kind of shell or e.g. of perl. It's enough, that your script containsshebangline. Therefore you should remove /bin/sh from your cron job line.Another aspect, that might cause a different behavior of your script by interactive start and by cron daemon start is...
I am on shared hosting and I'm trying to schedule cronjob to run every now and then. Via cPanel I scheduled to execute my script but even though that according to my host support the cronjob runs, the script doesn't seem as doing anything. The cron job command I set via cPanel is:/bin/sh /home1/myusername/public_html/s...
Simple script run via cronjob doesn't work but works from shell
I have ajavaapplication running inside a docker and i use:docker exec -it <containerName> jstack > someFile.txtJust make sure to collect it periodically to have a better understanding.Aslo, You can use applications likeyourkitto do it for you
My application running core OS and Ubuntu in a docker. How to collect thread dump from these systems to analyse the performance issuesi have tried below commands but it didn't work:kill -3 jstack docker exec appsnap
How to take threaddump from docker
SeeGCP's release statement:Istio is being developed in the open on GitHub, and we invite the community to join us in shaping the project as we work toward a1.0 release later this year.
I'm interesting to useIstiooverKubernetes. However, I found out that Istio still under development. (version < 1.0.0)When is the first stable version going to be release?Is it safe to use the current version?
When is Istio gonna be stable (version 1.0.0)?
Welcome to the community.Unfortunately there's no such list with allimmutable fieldscombined in one place.There are two options:As you started through reading documentation and see if this is specified explicitly.Start withkubernetes APIdescription. You can find it here:Kubernetes API. This is also available in more hu...
Is there a way to find the immutable fields in the workload's spec? I could see there are few fields mentioned as immutable in some of the workload resource documentation, But for example inStatefulSetis not clear which are immutable fields. Is there a better way to find out?Sorry, I am not so familiar with reading the...
How to find the immutable fields in a kubernetes workload spec?
The connection-name is the name of the github connection in your zuul.conf and can be chosen freely.
I want to connect Zuul CI with GitHub This page describes how to do it. The part that I do not understand is: Set Webhook URL to http://<zuul-hostname>:<port>/api/connection/<connection-name>/payload. What is <connection-name> here? I tried the tenant name here, but this did not work. Any hints?
Connect Zuul CI with GitHub
serialNois aDisplayString. Prometheus values are floats, not strings, so the way that the snmp exporter exposes strings is to do so as labels. This is a form of info metric, which are gauges andhave a value of 1.
I want to get the serialNO value, which is of type string, but snmp_exporter gets the gauge type, snmp.yml file as follows:- name: serialNo oid: 1.3.6.1.4.1.258.120.1000.6.2.2.2 type: DisplayString help: ' - 1.3.6.1.4.1.258.120.1000.6.2.2.2' indexes: - labelname: daNo type: gaugeWhen you look ...
snmp_exporter (prometheus) - get error type of metrics
These days you can use the experimental Docker feature checkpoint and restore:https://github.com/boucher/docker/blob/cr-combined/experimental/checkpoint_restore.md.There is a howto available athttps://criu.org/Docker.
I'm trying to find a way to perform a memory dump on a docker container in order to perform memory forensics (to detect malware exploits for example). I would like to be able to perform the same methods I use on a virtual machine. The problem is that docker containers (and any kind of linux containers) use memory in a ...
How to perform memory dump to docker container from outside
Create a rule to open port 3000 in the security group associated with your ec2 instance. It can be done through the command line tools or through the web console, which is more straightforward. If you didn't specify a security group when creating the instance it will be the "default" security group.A decent walkthrough...
I'm trying to run a basic node.js server,var http = require('http'); http.createServer(function(req, res) { res.writeHead(200, {'Content-Type': 'text/plain'}); res.end('hello world!\n'); }).listen(3000, '0.0.0.0', function() { console.log('Server running on port 3000'); });However when I run it and go tohttp://x...
I can't access my node.js server on my AWS EC2 isntance from the outside
A Docker container runs only as long as its initial command does. In this case, that command is service apache2 start, which exits after no more than a couple seconds, at which point the container stops as well. If you want to run a daemon in a container, you need to either make the daemon the primary program and ma...
I am trying to understand how docker works when it comes to starting / getting into a container. I am new to this so bear with me. I tried finding the answer to my need in SO but could not so far. I am trying to achieve 3 basic things: Start a container Make sure all ports are mapped to my container (i.e. I can see ...
Docker container doesn't start, showing as 'Exited n seconds ago'
You cannot add conditional policies that apply to objects in S3 object lifecycle configuration which in your case is, is based on the object's last access time.You can however transition objects to Glacier based on their age or on a specific date.I would like to think you can handle it in your application but the s3 ob...
i am storing objects to S3, i would like that object never accessed in the last month go to glacier.After some research i don't think i can achieve this, but i hope to be wrong.When creating lifecycle for an s3 bucket the rule is based on object creation date (not last access date)Setting the storage class for the obje...
AWS storage: S3 object go to Glacier if never accessed in the last month
No, I think you're pretty much right.The image you link to is an automated build, so OpenBlas will be getting compiled using the kernel and architecture of the Docker build server. I did notice thebuild scriptsets the following flags when building OpenBlas:DYNAMIC_ARCH=1 NO_AFFINITY=1 NUM_THREADS=32Which presumably mak...
Some libraries such as BLAS/LAPACK or certain optimisation libraries get optimised for the local machine architecture upon compilation time. Lets take OpenBlas as an example. There exist two ways to create a Docker container with OpenBlas:Use a Dockerfile in which you specify a git clone of the OpenBlas library togethe...
Docker images with architecture optimisation?