Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
Make sure that the pylintrc (or sometimes called .pylintrc) file is at the top level of your project. Theoretically pylint should be able to pick up that location by default, but if it still isn't working, pass in the argument --rcfile=pylintrc to pylint. For example: pylint --rcfile=pylintrc $(git ls-files '*.py') P...
Closed. This question needs debugging details. It is not currently accepting answers. Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help oth...
Github workflow pylint: How do I disable all docstring errors warnings? [closed]
You don't need a special function. Just do it like this double (*A)[n][m] = malloc(sizeof *A); As of C99, here n and m can be any positive integer expressions you want. Such a thing is a pointer to a VLA, variable length array.
I have a simple question, but the answer seems to be very difficult to find: How do I create a true 2D array in C (not C++), dynamically sized (size not known at compile time), not an array of pointers, on the heap, so that I can put that allocation into a separate function and return the allocated array, without rece...
Allocate 2D Array in C (not array of pointers) in Heap
You could place thepublish.cmdfile in your.gitignorefile. If you want to be the only responsible person for publishing and deploying the bot then I don't see any problem with proceeding this route.I don't work with C# Bots as often as I do with Node.js bots, but if you're doing continuous integration you should be rebu...
I recently created my first Microsoft Bot Framework bot and wanted to put it up on my public github so a few friends could add to it. What is the recommended way to do this? I see that they have Continuous deployment setup from github, so I'd imagine there is a recommended way to do this.I tried downloading the source ...
Putting Bot Framework bot on Github
+50AKubernetes objectis a persistent entities in the Kubernetes system.AKubernetes resourceis an endpoint in theKubernetes APIthat stores a collection ofAPI objectsof a certain kind; for example, the built-in pods resource contains a collection of Pod objects.ObjectsKubernetes uses these entities to represent the state...
What is the difference between Objects and Resouces in Kubernetes world?I couldn't find it fromhttps://kubernetes.io/docs/concepts/. I wonder they make no distinction about them but seems they see objects as a high-level concept of resources.
Difference between Kubernetes Objects and Resources
1 I posted this on the issue tracker and got this answer: There are a couple of things going on. The VM on older devices uses conservative collection. Most (but not all) devices running >= 2.0 will use type-precise GC, but none of them yet have live-precise GC. ...
I'm having some trouble understanding why this code public class BitmapAllocTest extends Activity { /** Called when the activity is first created. */ @Override public void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); byte[] b = new byte[20 * 1000 * 1000]...
Android bitmap allocation weirdness
You're using the wrong constructor. Just create a bitmap directly from the unmanaged data, and it will wrap the data, rather than creating a new bitmap just to throw it away immediately: new Bitmap(_size.Width, _size.Height, 4 * _size.Width, PixelFormat.Format32bppArgb, unmanagedByteDataPtr);
What is the most memory efficient way to redraw an Image with new data in C#? The external DLL I am using allocates a byte buffer to hold my (width * hight * bytesPerPixel) pixel data where each pixel is in ARGB32 format. The DLL automatically updates that buffer with new data by interacting with hardware, and then ca...
C# efficiently redraw Image with new data
1 Have you tried docker-compose up with -V option? -V, --renew-anon-volumes Recreate anonymous volumes instead of retrieving data from the previous containers. Share Follow ...
The docker-compose.yml contains a volume definition and a service that uses it. #(This is a reduced but representative example) services: mongo: image: mongo:3.6.11 ... volumes: - "mongodb:/data/db/" volumes: mongodb: driver: local driver_opts: type:...
How to force docker compose recreate manually deleted volumes?
Whatever is available to you will do just fine. When you runnpm run build staticit will create static files that you are able to host from whatever web-server you want. If you are curious about performance numbers, they are negligibly different in this case. Just choose whatever platform you feel most comfortable with.
I am new to reactjs. I created a basic react app usingcreate-react-appboilerplate. I prepared the build by running the commandnpm run build.Now I want to host the build folder. Which is better and why.Please let me know if any better and easy option available as I am new to this.
NGINX vs. Apache? What to choose when deploying react
Microsoft released a python docker image for PlaywrightDockerfile# Build Environment: Playwright FROM mcr.microsoft.com/playwright/python:v1.21.0-focal # Add python script to Docker COPY index.py / # Run Python script CMD [ "python", "index.py" ]CheckPlaywright - Docker docsfor the latest playwright version.
I need to install playwright inside of docker. This is my dockerfile.FROM python:3.9 EXPOSE 8000 WORKDIR /fastanalytics COPY /requirements.txt /fastanalytics/requirements.txt RUN pip install --no-cache-dir --upgrade -r /fastanalytics/requirements.txt RUN playwright install RUN playwright install-deps RUN apt-get u...
Docker and playwright
-1 Probably the easiest way to accomplish it is to disable the integrated graphics card. If you are only interested in your own computer, then you can disable it through the device manager. Access through right click on "my computer/manage" left side panel, Device Manager (...
I'm learning Three.js and created a simple app with it expecting it to run on my dedicated GPU, but task manager shows 0% GPU utilization (all values) while the scene is rendering and 80-90% integrated graphics usage. I tried a few sites that use WebGL and the results were the same. Here's what I know: Hardware accel...
Three.js/WebGL running on integrated GPU instead of dedicated
A large file suggests more work to be done. So valgrind needs more time. Valgrind really isveryslow.You can easily debug this with the world's best debugger: printf() (only half-kidding.) Simply print something before or after every iteration of your main loop. If it doesn't show up, valgrind is really hanging somewher...
For a method that I am trying to code for one of the classes I have been working on, I am trying to read double values from a file and dynamically set some arrays inside the program with these numeric values.I wanted to check, at least up to the point that I came, whether I have memory leaks or not. However, firing upv...
Valgrind is hanging with no output
I wrote to the support of cnvrg and adding machines is not allowed as CORE is a free community version the option to add new resources is not supported. Other paid versions of cnvrg allow you to add new resources and other functionalities, so cnvrg CORE is not a real open-source version of cnvrg.
Good afternoon,I have recently deployed cnvrg CORE application on-premise with Minikube.In cnvrg CORE we can create a "machine resource" to give to the application some computer resources like CPU and GPU from a different machine through SSH.I have found a problem when creating a new resource of any type (in the attach...
Can't add resources to cnvrg CORE
+25You could build you own image, using a Dockerfile.There are two ways :Take thecurrent image Dockerfileand add what you need. If you do this way, you can delete the unused packages too.Create a new one, and just specify that your image is based on the current one, usingFROMkeyword.
I am using this image that is working as expected.https://hub.docker.com/r/clue/adminer/The only problem is that it does not has php extensions required to connect MongoDB or Oracle...Oracle None of the supported PHP extensions (OCI8, PDO_OCI) are available. MongoDB None of the supported PHP extensions (mongo) are ava...
php support for mongoDB
4 Check if you have changed the "Components" from uppercase to lowercase "components". In my case, I changed it after it had been created on GitHub. But GitHub doesn't update this change. When it's cloned to Amplify host, it's still "Components", so the build failed. I ha...
for some time, I have been wrapping my head around an error preventing me from deploying my react app to AWS Amplify. The compiler seems not to be able to resolve a component during the build phase. I keep getting the following error Module not found: Error: Can't resolve './components/...' in '/codebuild/output/src31...
Amplify React build fail with "Module not found: Error: Can't resolve './components/...' in '/codebuild/output/src313005886/src/client-portal/src' "
You have to find the instance that supports your engine postgres and version 13.3 for your region. To do this you should use the following command: aws rds describe-orderable-db-instance-options --engine postgres --engine-version 13.3 --query "*[].{DBInstanceClass:DBInstanceClass,StorageType:StorageType}|[?Storage...
I am trying to create postgres rds using cloudformation but it give me this error RDS does not support creating a DB instance with the following combination: DBInstanceClass=db.t2.micro, Engine=postgres, EngineVersion=13.3, LicenseModel=postgresql-license. For supported combinations of instance class and database eng...
RDS does not support creating a DB instance
df command you need to run in the container and also go through thisdoc, /bin/df as a tool to monitor ephemeral storage usage on the volume where ephemeral container data is located. Refer to thisblogby Sudip Sengupta for more information.Try the below commands to know kubernetes container's ephemeral-storage usage det...
I know how to measure needed RAM and CPU usage, but not how to measure used disk space of every container in a k8s cluster.I've tried cAdvisor, but only provides stats per node, not per container.https://github.com/google/cadvisor/issues/198Any recommended tool/procedure? In this way I could specify in theresourcessect...
How to measure the ephemeral space used by every container
Here's the solution - When I go to my Plesk's Web server settings, there is a checkbox that says "Process PHP by Nginx". Check that, and the phpinfo(); shows the Server API as FPM/FastCGI now! Cheers!
I have just installed Plesk 12, which comes inbuilt with an nginx server over an apache server. On enabling PHP-FPM, i cannot see it mentioned using the phpinfo(); function...thought the website works just fine. Also, on trying these - 1.) telnet 127.0.0.1 9000 Result = telnet: could not resolve 127.0.0.1/:9000: Ser...
Check if PHP-FPM is working
In summary, you need to abort the build without failing.How to abort?However, you are only testing thewhencondition into one single stage, this is why the stage is skipped but the build continuous.As per design, thewhenexpression for Jenkins DSL cannot be executed outside a stage.DocumentationSo you have two alternativ...
I have a C# repo that has multiple Projects inside of a single solution. Each Project has their ownJenkinsfilebut they could share a Pipeline with another Project. Inside of theirJenkinsfileit gives a subdirectory. I am trying to get it to where it will only continue through the Pipeline if there have been changes to a...
Ending Jenkins Pipeline Early
CloudFront can't do this by default -- CloudFront-Viewer-Country is intended as a request header, sent to the origin, rather than a response header, sent to the browser. However... with a Lambda@Edge Origin Response trigger, it is possible to achieve what you appear to be trying to do: echo this header and its value b...
I am attempting to leverage AWS's geotracking feature on CloudFront to inform my UI of user location so that it knows not to load certain files in problematic regions. AWS documentation mentions the following but does not give clear instructions on how exactly to configure CloudFront to forward the CloudFront-Viewer-C...
How do I get CloudFront-Viewer-Country to appear in response headers?
All nginx URIs begin with a leading /, and your regular expressions do not attempt to remove the /api/v1/ prefix before appending the parameter list. Try: location /api/v1 { rewrite ^/api/v1(/.*)$ /api/v1/routes.php?params=$1 last; }
I'm performing a migration from an Apache server to Nginx and missed a .htaccess file that performs a rewrite: RewriteRule ^(.*)$ routes.php?params=$1 [NC,QSA] For example, this endpoint example.com/api/v1/param1/param2/param3 would be rewritten as https://example.com/api/v1/routes.php?params=/param1/param2/param3 Ca...
Nginx rewrite equivalent to Apache RewriteRule that converts URL params into QueryString key/value pair
You need to enter in the username and password for the remote url. I have tested the following in the past and it worked for me:git remote set-url origin https://$USERNAME:[email protected]/git/user/projectwhere USERNAME and PASSWORD are environment properties on the stage. I recommend setting the password as a secure ...
I have a Build & Deploy Pipeline in Bluemix, I would like to tag git if the stage has been successfully deployed. Currently, I had added a build step after a deploy step into my deploy stage with a shell script that look like:# put the git tag echo 'Put tag build_$BUILD_NUMBER on git' git tag build_$BUILD_NUMBER git pu...
Bluemix: How I can configure a delivery pipeline stage build to tag git?
Or you can put it simply in its own location -location /robots.txt { alias /Directory-containing-robots-file/robots.txt; }
I am running nginx 0.6.32 as a proxy front-end for couchdb. I have my robots.txt in the database, reachable ashttp://www.example.com/prod/_design/mydesign/robots.txt. I also have my sitemap.xml which is dynamically generated, on a similar url.I have tried the following config:server { listen 80; server_name example...
How do i configure nginx to redirect to a url for robots.txt & sitemap.xml
You can't do what you want in one entry, since the two minute definitions will apply for both hour definitions (as you've identified). The solution is (unfortunately) use two cron entries. One for 00:00 and one for 13:30. An alternative is perhaps to execute one script at 00:00. That script would execute your original...
i want to execute a script twice daily at 00:00 and 13:30 so i write : 0,30 0,13 * * * it seems wrong for me, because like this, the script will fire at 00:00 , 00:30 , 13:00 and 13:30. Any idea ?
execute crontab twice daily at 00h and 13:30
Whether to check the Private or Public profiles depends on how your network connections (more specifically the local area network) are labeled.You can check and alter the network type underControl Panel\Network and Internet\Network and Sharing Center.With the shown setup, at least partial access is allowed from connect...
I would like a Express REST API to be accessible over a LAN.From what I understand I will need to make some changes in my firewall to allow this access. I managed to solve this checking the second and last mention ofNode.js: Server-side Javascriptin Windows Defender.Looking at the supplied image I have a feeling that I...
Required firewall exceptions for accessing Node apps on a LAN
Answering my own question here for whoever might be learning something from it.github's diff api is what you're looking for.Like so:https://api.github.ibm.com/repos/{account}/{repo}/compare/${buildCommit}...${latestMasterCommit}
I'm on a older github enterprise server (3.2). Developer usually commit to develop, but some hot fixes can happen on master as well.Each time when we deploy, we generate a github release. Up until today, we used the "recent changes" feature from Jenkins (which builds from master) in order to generate a changelog, howe...
Getting recent changes for generating release notes on github enterprise
All following elements are moved one step back. Vector guarantees to hold its elements in continuous block of memory, so no internal holes are allowed.
Since the memory locations are contiguous for stl vector are rest of elements moved back by one step? How did the stl library writers implement it?
What happens in STL<vector> when an element is removed from between?
Well the problem was with postgress - you have to add your ip in pg_hba.conf => host all ip trust/md5(depend on version). And thanks to a_horse_with_no_name for editing it in proper format
My my_backup.rb=> database PostgreSQL do |db| db.name = "xxxxx" db.username = "postgres" db.password = "*********" db.host = "localhost" db.port = 5432 end store_with SCP do |server| server.username = "us...
Trying to have postgres database backup by backup rubygem nad ruby on rails
Quick answerThe DKIM specification (RFC 6376 Sec. 3.6.2.2) dictates thatStrings in a TXT RR MUST be concatenated together before use with no intervening whitespace.In other words: The whitespace in between the strings does not have any relevance. The strings inside the quotes are simply processed as one.Technical backg...
We've added a TXT record for DKIM validation (copy-pasted the DKIM string), but there seems to be a weird character in the record that:doesn't appear at all in the DNS managerdoesn't appear at all in theDKIM Corevalidatordo appear as empty quotes in themail-tester.comvalidatordo appear as a whitespace within quotes dur...
Weird whitespace in TXT record
I ended up creating php class to handlepre-package-installevent thrown by Composer.<?php namespace Company\Composer; use Composer\Installer\PackageEvent; require_once(__DIR__ . "/../../../vendor/autoload.php"); class EventHandler { public static function prePackageInstall(PackageEvent $event) { $ope...
I have a GIT repository that creates releases which can be downloaded and installed via Composer.Those releases currently create tarball containing entire repository. Problem is that repository is huge and I do not need entire contents.Aside from that there is a build process that compiles, merges etc. and provides a p...
Composer custom release package
The dependencies are resolved on a need-by-need basis: The compile time ones for the application's code are downloaded as a very first step when you start the build. Any plugins required are downloaded afterwards during the respective phases. This also includes their transitive dependencies. The dependencies for the ...
How can I warm up the dependency cache of my maven tests? E.g. mvn test -DskipTests downloads some of the dependencies, but not all, e.g. some of the maven surefire plugin dependencies are only downloaded by mvn test. I want to create a snapshot of my filesystem to boost up the execution of my tests. Therefore I'd lik...
Warm up maven dependency cache
UPDATE(2018-01-31):It looks like Cloud HTTP(S) Load BalancersupportsGZIP. You just have to serve compressed content from your backend and the load balancer will pass it on.However, NGINX is confused because of theViaheader (it thinks proxies don't support GZIP, and on most cloud providers this is correct, but not Googl...
My backend, in Rocket (Rust), does not have compression built in. So, it is dependent on the proxy to compress it. Though nginx ingress controller supports it, I thought whether the default one had it too as it has high availability.If it does not have, then how should I setup?
Does GKE Ingress-GCE support compression?
1 Does SharePoint 2013 restore only from Database? The short answer is no. A full fidelity SharePoint farm backup is mostly databases but there is also configuration information and data that is stored outside of the databases. The Central Admin backup facility (as well a...
Does SharePoint 2013 restore only from Database? I have a scheduled script in MSSQL Server to run all database backups daily , and my SharePoint site also require a daily differential/weekly full backup usually happen in Central Administration. I am aware that multiple backups running would break log chain in this ca...
SharePoint 2013 Backup and Restore from database side only
+100Did you check this SDK? Seems very recent but might do what you need.https://github.com/awslabs/amazon-transcribe-streaming-sdk/tree/masterIt looks like it handles the signing:https://github.com/awslabs/amazon-transcribe-streaming-sdk/blob/master/amazon_transcribe/signer.py
I am working with theAWS Transcribe streamingservice that boto3 does not support yet, so to make HTTP/2 requests, I need to manually setup theauthorizationheader with the "AWS Signature Version 4"I've found some exampleimplementation, but I was hoping to just call whatever function boto3/botocore have implemented using...
Generate the AWS HTTP signature from boto3
The key you are trying to get doesn't exist among your environment variables. Changing the code to - os.environ.get('TWITTER_ACCESS_TOKEN') or any other key among your env vars should do the trick.
Recently I have been trying to deploy a django webapp to AWS Elastic Beanstalk and everything has been going fine. However part of my app uses that Twitter API so I need to import my API keys. My understanding is that I should use Configuration > Software Configurations > Environment Properties. I set this up inputtin...
AWS Elastic Beanstalk Environment Variables in Python
Yourcronfile and scripts seems correct. There some things to test though.First, make sure you script has execution permission (specially forpostgresuser):$ chmod +x /var/lib/pgsql/test.shI'd also set it topostgresuser as owner:$ chown postgres:postgres /var/lib/pgsql/test.shNow. Make sure thepsqlbinary is mapped in you...
I created a file called 'test_purge' in /etc/cron.d that contains:00 04 * * * postgres /var/lib/pgsql/test.shThis is supposed to run test.sh at 4 AM everyday, right?test.sh has:#!/bin/bash psql -d kettle -c "TRUNCATE TABLE test;"Am I supposed to do anything else to make this run? Do I add 'test_purge' which is in cron....
How do I run a cronjob from cron.d?
Will it's simplest form passing only Bucket, Key and SourceFile attend my goals?The answer isYesit will serve your purpose but if you use Metadata then you can have more control over your object.According to AWS documentation about Object Metadata,There are two kinds of Meta data:System metadata: Metadata such as objec...
InAWS SDK for PHP v3the methodputObjectcan receive many parameters likeContentType,ContentEncoding, etc.In it's simplest form, I can put a object using onlyBucket,KeyandSourceFile:$result = $s3->putObject(array( 'Bucket' => $bucket, 'Key' => $keyname, 'SourceFile' => $filepath ));Considerin...
Why should I use metadata when putting objects in AWS S3?
Heroku uses git for deploy. So you can use it as version control too.But I would not recommend it. When you push to heroku it's mean deploy to production. But your code can be not ready for it. Not tested yet, feature not fully implemented and etc.You can add 2 remote for your repository.git push origin master # github...
I am working with this school project (webapp in RoR) in group of 10 and we get into this fight.One says we should use Heroku as our web host because it does version control with git. The other says it's cool to use Heroku as web host, but it doesn't not store old code and keep track of changes, so we should set up ou...
Can Heroku work as version control?
There is no need to synchronize beforecudaUnbindTexture. It contains implicit synchronization.
I'm using a texture memory for one of my kernel programs.I bind the texture memory, than call the kernel execution that uses the texture memory, and finally I unbind the texture. I'm trying to figure out if there is any possibility that my device would unbind texture before my kernel execution end? Or either way, woul...
Should I synchronize before unbinding a texture in CUDA?
There are two possibilities here.Not Packaged, Not FoundPackaged, Not FoundGo into the.jarpackage, and check if yourmock/x.jsonexists intarget\example.jar!\BOOT-INF\classes.If it is being packaged but not found, seethis answer.new ClassPathResource("./mock/fileName.json")and then useresource.getInputStream().I strongly...
In my SpringBoot application I have to read a Json file at this path src/main/resources/mock/fileName.json. I done it in this wayJsonReader jsonReaderStream = new JsonReader(new InputStreamReader(Objects.requireNonNull(ClassLoader.getSystemClassLoader().getResourceAsStream("./mock/fileName.json"))));It works locally bu...
How to read a Json file under the resources folder?
Yo might want to use thecronmodule.- cron: name: "Run script.ksh Mo-Fr at 8AM" minute: "0" hour: "8" weekday: "1-5" job: "/appli/script.ksh > /var/log/cronlog.log 2>&1"ShareFollowansweredOct 29, 2018 at 15:00Vladimir BotkaVladimir Botka63.3k44 gold badges3939 silver badges6868 bronze badges12Suggest...
I would like to use Ansible in order to edit a crontab as the following line :00 08 * * 1-5 /appli/script.ksh > /var/log/cronlog.log 2>&1Any ideas ? Thanks !
Edit a crontab with Ansible
No. It takes 32 bits to reference a contiguous region of any size. If you have a 1 megabyte buffer, you're not going to store a pointer to every byte inside it, you'll just store a pointer to the beginning of it.
In a byte addressed space with 32bit addressing, it takes up 32bits of memory to reference 8 bits? So the addressing is the major portion of Memory? Am I conceptualizing this correctly?
In a byte addressed space with 32bit addressing, it takes up 32bits of memory to reference 8 bits?
<div class="s-prose js-post-body" itemprop="text"> <p>To run multiple commands in docker, use <code>/bin/bash -c</code> and semicolon <code>;</code></p> <pre><code>docker run image_name /bin/bash -c "cd /path/to/somewhere; python a.py" </code></pre> <p>In case we need command2 (python) will be executed if and only if c...
<div class="s-prose js-post-body" itemprop="text"> <p>I'm trying to run MULTIPLE commands like this.</p> <pre><code>docker run image cd /path/to/somewhere &amp;&amp; python a.py </code></pre> <p>But this gives me "No such file or directory" error because it is interpreted as...</p> <pre><code>"docker run image cd /path...
docker run <IMAGE> <MULTIPLE COMMANDS>
What you are describing is not possible. Only available operations for labels are=,!=,=~,!~.You can try a workaround based on regex selector. For example you provided in the question you'll need querynode_signal{date=~"0[1-7]\d\d|0801"}But this approach is not pretty: it requires new regex for every new day and those r...
For example, my metric looks like this:node_signal{date="0731"} 1 node_signal{date="0801"} 1.5 node_signal{date="0802"} 0I'm looking for something to query like this in grafana:node_signal{date <= "0801"}So it returns the first and the second metric. I have no idea is it possible to do this with promql in grafana panel...
Can PromQL query metrics that some label less equal than some value?
A PR has nothing to do withgit, really; that's something that GitHub has created to ease the process of collaborating with others. That means that thegitcommand itself doesn't know anything about pull requests.You have a few options if you want to interact with pull requests on the command line.GitHubprovides instructi...
Is there any git Command to get the commit message for a specific PR.git loggives me the complete history, but I have a PR and wanted to list all the commit message for the PR.
GIT Commit Message for a PR
Sure you can... You can simply create a new file inside of a repo. EDIT: According to this blog post, there is no way to directly upload files anymore. This feature was deprecated as of December 2012.
I need to upload a file to my repo. Can I do this using, just the web interface? Here is a related question w/ no answer, just link outs. How To Upload Files on GitHub
Is it possible to upload files to a git-hub account / repo with out downloading the application?
Before the changes are externally visible, you need to Deploy the API !
When executing a testvia the AWS API Gateway console, I see that the output response from the Lambda function is correctly transformed: { "type" : "", "message" : "", "request-id" : "" } See logs below: Tue Sep 06 14:46:06 UTC 2016 : Endpoint request body after transformations: {} Tue Sep 06 14:46:06 UTC 2016 :...
Amazon API Gateway : response body is not transformed when the API is called via Postman?
No, you cannot tell the origin repo to "undo" the last change pushed remotely. YoucouldSSH in and do agit reset --hard [revision]if you have access, though. However…If you didn't--forcethe push, Git didn't overwrite anything on the origin repository. In fact, if you pushed and git didn't refuse it, that means you were ...
Ok, so I work on a project from home and work using 2 computers with a remote repo.When I got home today, my home computer was 2 days behind my work computer and I did agit push origin masteraccidentally which has screwed up the remote repo (newly cloned code is now broken).I've seen the accepted answer fromthis questi...
Undoing a git push from a different computer
0 Changing the upload limit size takes just 5 steps Go to Plugin editor from the left menu panel in the admin dashboard. Top right, Choose All-in-one WP Migration from the dropdown and click select. Click on constants.php file. Search for the word AI1WM_MAX_FILE_SIZE. C...
Import 169,7 MB wordpress backup, but not running, if i refresh page but i see error, Does this happen if there is a difference in version? uninstall mysql db and user then create new db & user installing wordpress
Wordpress backup import using all-in-one plugin not working
I found an undocumented featuresymfony 2.8, php 5.6, swiftmailer-bundle 2.5.3Swiftmailer Configurationswiftmailer: stream_options: ssl: verify_peer: false verify_peer_name: false
I need to disable the validation of ssl certificate for developing purpose but i don't find anything about this in official documentation.http://swiftmailer.org/docs/introduction.htmlI'm using php 5.6 and Symfony2 (v2.7).The configuration reference of SwiftMailerBundle is:swiftmailer: transport: smtp u...
SwiftMailer - PHP - How to disable ssl certificate validation
No, in-order execution pipelines can let instructionsfinishexecution out of order after starting in-order (especially loads are commonly allowed to do this, letting static instruction scheduling help hide load latency). All of this is possible without a ROB. Just scoreboarding register writes is enough to enable that...
Closed.This question does not meetStack Overflow guidelines. It is not currently accepting answers.This question does not appear to be about programming within the scope defined in thehelp center.Closed4 years ago.Improve this questionWe know that a ROB exists in CPUs with out-of-order pipelines to reorder u-instructio...
Does a ROB exist in CPUs with in-order pipe-line? [closed]
You can define an environment variable no_proxy in order to avoid the proxy for a specific domain:no_proxy=.mycompanyThe other approach,since git 1.8.5+, is to set a proxy per url.
I use git behind a firewall and I need to configure$ git config --global http.proxyto handle connection.But what happens if I only wanna avoid proxy for local domains? I need the proxy for every host excepts for servers in my LAN.There is a way to do?Thanks
Exclude hosts or domains with git proxy
If you have a "generic"runtime.Object, you can use thedynamic clientin client-go for this. The dynamic client deals withunstructured.Unstructuredobjects and allruntime.Objects can be converted to it. Here is an example:// create the dynamic client from kubeconfig dynamicClient, err := dynamic.NewForConfig(kubeconfig) i...
I'm using AWS' EKS which is Kubernetes v1.10 and I'm using client-go v7.0.0.What I'm trying to do is parse a .yml file with multiple Kubernetes resource definitions in a file and submit those resources to the Kubernetes API. I can successfully parse the files using this codescheme.Codecs.UniversalDeserializer().Decode,...
How to Submit generic "runtime.Object" to Kubernetes API using client-go
The LoadBalancer should be getting created automatically.There might be IAM policy issues preventing the load balancer from being created (seeIssue #10692).If that isn't the problem, looking for errors in/var/log/kube-controller-manager.logon the master VM may give you an idea of what is going wrong.
i try to use AWS to setup kubernetes(version 1.0.1) and deploy a few services and pods there.but i have got stuck with LoadBalancer service. According to the docs i just need to setup correct type of service and open ports in firewallBut service doesn't receive external IP. (ingress is empty)Do i need to create LoadBal...
how to setup loaderbalancer service of kubernetes in AWS
First of all, you need to install Prometheus and Grafana in your Kubernetes cluster following the instructions given for each:Prometheus:https://prometheus.io/docs/prometheus/latest/installation/Grafana:https://grafana.com/docs/grafana/latest/installation/Next, you need to understand that Prometheus is a pull-based met...
I have to build a monitoring solution using Prometheus and Graphana for a service which is built using React(front end)+ Node js + Db2(containerised) . I have no idea where to start,can someone suggest me the resources where to learn?Thank you.
How to monitor a container running db2 image using Prometheus and also react app using Prometheus?
4 You should use the tag --deploy-mode cluster that will allow you to deploy multiple executions to your cluster. That will make yarn handle the resources and the queues for you. The full example: spark-submit \ --class org.apache.spark.examples.SparkPi \ --master yar...
Is it possible to submit and run Spark jobs concurrently in the same AWS EMR cluster ? If yes then could you please elaborate ?
How do I run Spark jobs concurrently in the same AWS EMR cluster ?
Try this works with windows 10$vpnName = "YOUR_VPN_NAME"; $vpn = Get-VpnConnection -Name $vpnName; if($vpn.ConnectionStatus -eq "Disconnected"){ rasdial $vpnName; }
I'd like my Windows to connect to the VPN server as soon as it loads. How can I do it using Powershell?
Connect to VPN by Powershell
Your PATH environment variable ($env:PATH) is misconfigured: It contains an extraneous, unpaired " as part of its value: # Your $env:Path value - note the unexpected, unpaired " at the end. ...;C:\Users\JHdJ\AppData\Local\GitHubDesktop\app-1.6.5\resources\app\git\cmd" cmd.exe happens to ignore the extraneous ", but P...
I installed GitHub and Git, and want to add Git to the PowerShell path. I added it to the system path, it works in cmd.exe, and shows the correct path when I print $env:path, yet PowerShell still doesn't recognize git. I tried to restart, thinking it was maybe behind or something, but that didn't fix the problem. If ...
Why does PowerShell not seem to honour the $env:path variable?
From a technical standpoint, the end result is the same. Rebase is just a smart tool to do exactly what you describe.I've resorted to manual editing a few times in the past, but generally it pays off to get proficient at using interactive rebase using Vim or Emacs.rebase -iseems arcane at first, especially as the defau...
I have to rebase and then, squash and merge my changes. But, the rebasing is taking too much time and effort. So, what if, instead of rebasing, I delete the feature branch, create a fresh one from master and then, apply my changes at the top and push in a single commit?Will it be any different than rebasing?Will the gi...
Git rebase vs recreating branch and applying changes
I am afraid it's not possible.Please keep in mind that there are some differences between On-Premise Kubernetes cluster and GKE cluster. Most important is that GKE master is managed completely by Google and you cannot reach it or change anything there. For example inVulnerability and patch managementdocumentation you c...
I have created a k8s cluster in GKE. But I want to configure API server for k8s audit purposes so I have to set--audit-policy-fileflag and--audit-webhook-config-fileflags as arguments in the API server. How do I do that?
Setting kubernetes audit policies within Google Kubernetes Engine
The easiest way is to ask them toforkthe repo and makePR (pull request).You can then inspect and/or reject any PR made to the master branch.TheGitHub branch protectiondoesn't prevent fast-forward pushes, to the fork is the best option.
I would like to give merge access to my co-workers. But, I do not want them to be able to merge their code into master (or another branch I maintain: staging). Is there any way to do that?
Is there any way to give Github merge access to people for all branches except master?
I believe you're running into trouble because the default shell run by Docker is not a login shell according tothis answer, which means scripts in/etc/profile.d/don't get processed.If you need profile processing, try changing your last line toCMD ["/bin/sh", "-l", "-c", "php71-php-fpm"]to invoke a login shell.
I am working in a Dockerfile for PHP-FPM 7.1. I am ending the Dockerfile with the following line:CMD ["php71-php-fpm"]Because I am usingdocker-composethis is how I start up the container:docker-compose up -dThe container compiles fine (apparently) as per this lines:Successfully built 014e24455b53 WARNING: Image for ser...
Executable file not found in $PATH
+25TheCAP theoremis vital for all large distributed systems. It is explainedhere.ShareFollowansweredSep 23, 2020 at 17:03Ari FordshamAri Fordsham2,45488 silver badges2929 bronze badges1Link-only answers are discouraged. Without those off-site links, your answer is just the phrase "CAP theorem" as something to google o...
Assuming you want to launch a social app (which meanmany interactions) with the ambition of acquiring several thousand users and for those who have already done so, what are the pitfalls that you know and that you would absolutely avoid, in term of code and servers architectures for sure ?I have that feeling that you c...
Code and servers pitfalls to avoid when launching social application with many interactions?
This is how I did finally: Igit initin each repo. Then I created new branches respective to that repo. When I moved a repo back and didgit branch, it only showed branches present within that current repo.Yes, but it also registered agitlink(aspecial entry in the index) for the nested repo... without registering the nes...
I have several repos within repos.In repo3 I want to create several branches. When I move back a repo, i.e. in repo2, and dogit branch, it shows me all the branches present in both repo2 and repo3. But what I want is to be able to create branches unique to their repos and and these branches shouldn't show up in repo2.T...
How to create unique branches in repos within repos
Answer recommended byCI/CDCollectiveThe actions runner is just a PC that clones the repo and runs a few commands in it. So when you create a new file on the runner, you have to add it to the repo on the runner, create a commit and push that to the repo over on GitHub.It might help to understand that the repo in GitHub ...
I'm new to GitHub actions, I want to use PowerShell and save the output of a command to a file in my repository, but I don't know what's the right way to do it.name: get hash on: [workflow_dispatch] jobs: build: name: Run Script runs-on: windows-latest steps: - uses: actions/checkout@v3 ...
How to create and save files using PowerShell in GitHub action?
I fixed it by adding.encode("utf-8")tosoup.That means thatprint(soup)becomesprint(soup.encode("utf-8")).ShareFolloweditedDec 4, 2019 at 19:36twasbrillig17.9k99 gold badges4343 silver badges6767 bronze badgesansweredNov 23, 2014 at 19:19SstrykerRSstrykerR8,34244 gold badges1313 silver badges1111 bronze badges47don't har...
I'm trying to scrape a website, but it gives me an error.I'm using the following code:import urllib.request from bs4 import BeautifulSoup get = urllib.request.urlopen("https://www.website.com/") html = get.read() soup = BeautifulSoup(html)And I'm getting the following error:File "C:\Python34\lib\encodings\cp1252.py",...
I keep getting 'charmap' codec can't encode characters error when trying to save python script's output to clipboard or text file [duplicate]
Kingfisher does not support server cache mechanism. It just uses the whole URL as the local cache key. As long as you use the same URL, you will get the same image from the cache (if it was cached). So, if your server provides different images under the same URL, we suggest you to ask your server to add a query to the...
I am using Kingfisher for loading image view with images from Url. Sometimes, the same url will updated with new image. So I am using below code to load the imageview, profileImage.kf.setImage(with: profileUrl, placeholder: #imageLiteral(resourceName: "profile_1"), options: [.fromMemoryCacheOrRefresh], progressBlock: ...
Show updated images in the same url using Kingfisher ios
With nginx, you have to concatenate the certificate chain and the cert into one file, in this case at/etc/nginx/certs/mitca.pemThis can be done on a UNIX CLI with the following command:cat your-signed.crt ca-cert-chain.crt > /etc/nginx/certs/mitca.pem...then restart nginx. You can usenginx -tbeforehand to make sure tha...
I run a webapp with users at MIT. They have asked whether the webapp could authenticate users (perhaps as a second factor) by their MIT personal certificates. I have tried but I can't get it to work.I downloaded the MIT CA certificate from theMIT certificates page– see the link "Get MIT CA (Certificate Authority)" ne...
How to get Nginx to verify MIT personal certificates?
When running inside a Docker container, by default not all of Nvidia libraries are mounted inside the container. Read morehereandhere. You can see all libraries available in your Docker container by running this:ldconfig -p | grep -E 'libcuda|libnvidia'To fix this, you should modify theNVIDIA_DRIVER_CAPABILITIESenviron...
I installedNvidia Video Codec SDK 8.2 + ffmpeginside adocker container by using nvidia-dockerbut when I run thisffmpeg -f rawvideo -s:v 1920x1080 -r 30 -pix_fmt yuv420p -i HeavyHand_1080p.yuv -c:v h264_nvenc -preset slow -cq 10 -bf 2 -g 150 output.mp4I got thiserrorCannot load libnvidia-encode.so.1The minimum required...
running ffmpeg with nvenc inside nvidia docker
1 quite an old question already but here are my 2 cents: invalidating a cookie is somehow just a set-cookie with an expired date, so I don't think you need to be able to read the cookie from the request to invalidate it. I have not tried but I would just do something lik...
This may sound like a very trivial doubt, but I need some help here. I have set up Application load balancer with OIDC authentication. After logging into my application, using chrome developer tool I looked into the client application cookie. Found that AWSELBAuthSessionCookie-0 and AWSELBAuthSessionCookie-1 have an e...
Application load balancer session cookie's expiration is extremely high
4 Looks like the static files don't have expires set. Read - http://www.absolutelytech.com/2010/08/02/howto-add-expire-headers-to-cache-static-files-using-htaccess/ You'll need to post the below code in your .htaccess # Turn on the Expires engine ExpiresActive On # Expir...
I got this error from Google Speed test: The following cacheable resources have a short freshness lifetime. Specify an expiration at least one week in the future for the following resources: http://localhost/english/favicon.ico (expiration not specified) http://localhost/english/images/bg_center.png (expiration not sp...
Google Speed Leverage browser caching
According tothisAdditional fields may be added in the future.If your controller is only ever writing to status then you don't need to bump the CRD version.It also depends what sort of validation you have on the CRD - like if thestructural schemais validating the status.Hereare conventions for multiple API versions (alt...
I intend to add additional error message fields to a custom resource status. The fields are marked asErrMsg string `json:"errmsg,omitempty"`Does adding these fields require a new API version, or is is safe to just change the managing operator to fill these fields with values?
Kubernetes: does extending the status subresource require a new API version?
I'm afraid there's a problem with that "newer" concept: in Git, "tree objects" (which record names, sizes and hashes of blobs they reference) do not contain timestamps of those blobs. Hence there's a logical problem of deciding which of two (or more, for octopus merge) blobs which happen to map to the same filesystem ...
i was just trying to merge two repositories. For some reason, the merging didn't go well and created a lot of conflicts. Most of them are due to the fact that i upadated a component in one version.So what i would like to do is merge and wherever a conflict pops up, tell git to use the newer version. Otherwise it would ...
Conflicts while merging - how to tell git to always use newer?
4 Try adding $context.error.validationErrorString and $context.integrationErrorMessage to the logs for the stage. I added a bunch of stuff to the Log Format section, like this: { "requestId":"$context.requestId", "ip": "$context.identity.sourceIp", "requestTime":"$context...
I created an API by AWS API Gateway and Lambda that is same 'https://github.com/aws-samples/simple-websockets-chat-app'. But the API not working trust. I get an error when i try to connect. Its message is "WebSocket connection to 'wss://b91xftxta9.execute-api.eu-west-1.amazonaws.com/dev' failed: Error during WebSocket...
AWS API Gateway WebSocket Connection Error
The increased CPU usage comes if you do not hit a pre-calculated entry. With LOOKUP 2 you are calculating 0-1023, but only storing 0, 2, 4, etc... So if you need the data for scratch-pad entry 3 you have to calculate it on the fly using the data from 2. This is an extra calculation vs. having them all stored permanent...
I'm developing a Litecoin Miner for a processor that has only 32KB of internal memory. So I was looking at SCrypt algorithms and for Litecoin it uses N = 1024, that gives me 2^10 * 1 * 128 = 128KB memory use aproximate. So I was looking into GPU Algorithms that has the parameter Lookup Gap. For reading I'm using keple...
SCrypt Lookup Gap Negative Effect
8 A queue is not a database table; you can't just 'read all the messages'. You tell SQS you want some messages, it picks some and sends them to you; you can't ask it, or expect it, to iterate thru a unique list of messages in the queue and present them to you, it is not i...
I want to read all messages from queue in AWS SQS without deleting them , using C# , but I am only getting the single last message. Here is the code ReceiveMessageRequest rmr = new ReceiveMessageRequest(); rmr.QueueUrl = url; rmr.MaxNumberOfMessages = 10; AmazonSQSClient sq = new AmazonSQSCli...
C# AWS SQS Read all messages without deleting
This answersets out the basic principle of using^<remote>/<branch>to exclude all commits reachable from the remote branch.This script should get you going:# See https://stackoverflow.com/a/18359250/3150057 for awk branch parsing. # Get all local branch names. LOCAL_HEADS=$(git branch | awk -F ' +' '! /\(no branch\)/ {p...
I'd like to back up (via dropbox or google drive or similar) the unpushed commits in my local repository. Backing up the entire repository (withgit push --mirroror similar) is redundant, since 99% of its commits exist in the central shared repo, as well as on all the other developers' machines on our team. Also, a fu...
Backup only "local" commits / objects in git repository
The above regex matches any file whose extension begins with j or J, and whose filename doesn't consist entirely oft'sh'su'sm's andb's. Not exactly what they were going for I don't think.RewriteRule ^(.*\.(?:jpe?g|bmp|gif|png|tiff?))$ watermark.php?src=$1will rewrite anything with the extension:jpegjpgbmpgifpngtifftif...
I have an auto-watermarking script on my website, and the .htaccess file uses this line to redirect all images to the watermark.php file, which watermarks them. That I all have working, except the .htaccess line also redirects things like .js. So now my jQuery doesn't work anymore. The line is this:RewriteRule ^([^t...
.htaccess Regex Watermarking
This problem could be either caused by 1. server configuration or by your 2. python code.for point 2, to be sure to exclude your code from the error, try this:import MySQLdb def dbconnect(): try: db = MySQLdb.connect( host='localhost', user='root', passwd='XXX', ...
I have a script that connects to a MySQL DB viapyMySQL.It works like a charm when I execute it manually from the console, but gives this output when I run this cronjob:@reboot sudo python3 /var/www/html/ls/src/AppBundle/Command/crawl.py true > /tmp/listener.log 2>&1Result:Traceback (most recent call last): File "/usr...
PyMySQL does not connect from Cron but from Console
Solved it, the issue was within my understanding of how GCP operates.To solve the issue I had to add a firewall rule which allowed all incoming traffic from internal GKE network. In my case it was10.52.0.0/24address range.Hope it helps someone.
I am trying to setup connection to my databases which reside outside of GKE cluster from within the cluster.I have read various tutorials includinghttps://cloud.google.com/blog/products/gcp/kubernetes-best-practices-mapping-external-servicesand multiple SO questions though the problem persists.Here is an example config...
Expose an external service to be accessible from within the cluster
you need to get familiar with the syntax for crontab entries. That syntax is:* * * * * command ┬ ┬ ┬ ┬ ┬ │ │ │ │ │ │ │ │ │ └──── day of week (0-7, sunday is 0 or 7) │ │ │ └────── month (1-12) │ │ └──────── day (1-31) │ └────────── hour (0-23) └──────────── minute (0-59)if you set a value at any place that fixes that pa...
I discovered a really weird behaviour in my TYPO3 scheduler: I created a task with the frequency* */1 * * *, so it should run every hour. But in "next execution" it's written, that it will run the next minute. So I tried to use "real" seconds, and set the frequency to "3600" (= 1 hour). Now it shows that the next execu...
TYPO3: wrong cronjob timings
The 5.1.2 issues search web service includes parameters which let you query for issues by creation date. Your best best is to use AJAX requests to get the data you need and build your widget from there.Note that you can query iteratively across a date range using&p=1&ps=1(page=1 and page size=1) to limit the volume of ...
I want to display a graphical report based on time (weekly/daily) which shows that what is the status of static code analysis over the period of time. E.g. vertical bar will denote number of issue and horizontal will display the time day/month/week. This will help to keep an watch of code quality easily over the period...
Sonarqube report in graph/chart for time (weekly/daily) and number of issues
2 There's plenty of workflows available, each defining different ways of using the git remote and local repos. To name a few: centralized workflow, feature branch workflow, gitflow workflow, forking workflow. Generally, remember that whatever's in your local repo, it is ...
Closed. This question is opinion-based. It is not currently accepting answers. Want to improve this question? Update the question so it can be answered with facts and citations by editing this post. Closed 7 years ago. ...
When should local repo be commited to and remote not? [closed]
Usually Gradle retrieves dependencies on demand, only if and when they are need. To make Gradle download all dependencies beforehand (and thus populate the local cache), you can use a task liketask resolveAllDependencies { description "Resolves all transitive dependencies (e.g. to build offline later)." doLast...
I am doing nightly gradle builds on a server where I checkout the repositories from git and build them.However, there is a proxy where gradle cannot download any repository. I tried running gradlew offline mode then I get an error telling me that"No cached version of" and then the name of the dependency.This is obvious...
Gradle offline how to cache the dependencies
4 I think the short answer is that Next Fit allocates from blocks throughout the whole free memory region, which means that all blocks are slowly reduced in size. First Fit allocates from as close to the front as possible, so the small blocks concentrate there. Thus the sup...
I'm reading the Chapter 21 Understanding the Garbage Collector of Real World OCaml. In the section Memory Allocation Strategies, it says: First-fit allocation If your program allocates values of many varied sizes, you may sometimes find that your free list becomes fragmented. In this situation, the GC is forced to pe...
How first-fit allocation algorithm reduce memory fragmentation?
git checkout -b staging upstream/stagingcreates a new branch named "staging" starting at the "upstream/staging" point - butgit checkoutdoes not pull or fetch. Have you rungit pullfrom there?
tried thisgit checkout -b staging upstream/stagingbut I can't get the source code of the upstream's staging branch.Do I need to reset or I've done something wrong?What I want is a clean copy of upstream/staging.
checkout from upstream failed to get the correct source
Not possible yet. It has been requested on the forums but no ETA. According to the Release History of AWS CloudFormation, the feature was added on Nov 19, 2018. This should replicate the fixed response you shown with the console pictures. MyServicesLoadBalancerListener: Type: AWS::ElasticLoadBalancingV2::Listener ...
From the Load Balancer Lister configuration page, in the AWS Console it allows you to create a listener with a Default Action as shown here: The Fixed Response option allows you to specify an http return code and a body: Below is sample known-valid CloudFormation. Not sure how to edit this to support non-forwarding ...
How to create a AWS LB Listener from CloudFormation that returns a fixed response?
You need to assign therole to lambdafunction to read from the secret manager.AWS roleThe following IAM policy allows read access to all resources that you create in AWS Secrets Manager. This policy applies to resources that you have created already and all resources that you create in the future.{ "Versio...
created a secret manager key (non-rotational)with plain text option and encrypted. When i tried to get the value in lambda function , I am getting the error as permission denied. Could you please help how to resolve the issue
AWS lambda function to use secret manager
This happens because you don't have sufficient permissions to access Docker. There are two solutions: Run the command with sudo. (not recommended) Add current user to the docker group using this command: $ sudo usermod -aG docker $USER $ newgrp docker Logout and log back in once you run this command and try runni...
I am executing following code to get list of all container using npm dockerode on Ubuntu 14.04 machine. Docker container are running properly. var Docker = require('dockerode'); var docker = new Docker({socketPath: '/var/run/docker.sock'}); docker.listContainers({all: true}, function(err, containers) { console.log('e...
After executing following code of dockerode npm getting error "connect EACCES /var/run/docker.sock" on ubuntu 14.04
This works for me after a merge request:export LATESTLOG=$(git log -1 --pretty=%B) export SOURCEBRANCH=$(echo $LATESTLOG | awk -F '\\\\n' '{print $1}' | awk -F '/' '{print $2}' | awk '{print $1}')
I have a GitHub action workflow where I am doing a merge of various different source branches into target branch main.Here is a test workflow to read the values of env variables:name: Print Github Env Variables on: push: branches: - main - test pull_request: branches: - main - test ...
Determine source branch name on merge push in GitHub action workflow
I suppose there could be a number of reasons for this to occur, but in my case it was some issue on our VPN. I encountered the issue again today, hopped off the VPN and pushed again, and it immediately worked.ShareFollowansweredMar 12, 2021 at 18:14Aaron LozierAaron Lozier29011 gold badge44 silver badges1414 bronze bad...
I have encountered this problem recently and I have never been able to find a coherent explanation or solution.When attempting to pusha specific branchto github, I am receiving the following error:client_loop: send disconnect: Connection reset by peer fatal: the remote end hung up unexpectedly fatal: the remote end hun...
git push - client_loop: send disconnect: Connection reset by peer
SmartGit is somewhat sensitive to ungraceful shutdowns which typically occur when rebooting your machine without having exited SmartGit before. To solve this, readd the repository to SmartGit's known repositories list usingRepository|Add or Create, then exit SmartGit usingRepository|Exitand restart.
Still in local repos folder they appear to be.What I have:Ubuntu18.04, attachedbitbacketrepos(they appear every run ofSmartGit)What I did: Addedgpgkey toSmartGit, successfully attached log/pass ofGitHubtoSmartGit. Same mail onSmartGitandGitHub.Everything works just fine, until I close app and rebootLinux. Then they're ...
SmartGit removes(from rep.panel) my github's cloned repos
Emailed Prerender with the same question and the answer from Todd at Prerender was, You should set up the token wherever you set up your middleware. As far as which middleware to use, the nginx middleware lets nginx handle the proxy in an asynchronous manner so that requests from crawlers are never even sent to your ...
Prerender.io allows you to install it by either editing the Nginx config or adding a Django middleware. What are the performace differences, if any, and which setup is better?
Install Prerender in Nginx or Django?
Unlike when setting your commit email address, it is necessary to use the full noreply email address that GitHub provides. You can find it under the "Keep my email addresses private" section on the GitHub email settings page. This email address should look something like: [email protected] Use this email address to ...
When I try to add a GPG key that uses my GitHub provided commit email address, it gives me an error that the email address is "Unverified." How should I fix this? I followed the guide on GitHub for generating a GPG key and adding it to your account, which says: Note: When asked to enter your email address, ensure tha...
How do I verify a GPG key with my commit email address on GitHub?
When I run the image built from the Dockerfile via docker-compose, the container exit before the command in docker compose run. That is expected. That command in the docker compose file : s3cmd ls is appended to the entrypoint of your s3' Dockerfile. So something like that is executed : /s3cmd_repo/entrypoint.sh ...
Im trying to write a docker image for s3cmd. When I run the image built from the Dockerfile via docker-compose, the container exit before the command in docker compose run. Dockerfile FROM alpine:latest COPY ./.s3cfg /s3cmd_repo/.s3cfg.example COPY entrypoint.sh /s3cmd_repo/ RUN chmod +x /s3cmd_repo/entrypoint.sh E...
Docker exit after entrypoint
You can:get the default branch usinghttps://api.github.com/repos/octokit/octokit.rbcompare the specified branch to the default branch usingcompare two commits APIand extractahead_by&behind_byfields.In that case it would be :https://api.github.com/repos/octokit/octokit.rb/compare/kytrinyx/generator/spike...masterExample...
Using the Github API (no local git commands), is it possible to compare a branch to see if it has any changes ahead of the default branch?I'm building an auditing tool, and would like to identify branches that are candidates to be closed, because all of their changes exist in the default branch.I want the same informat...
Using the Github API is it possible to determine if a branch is ahead of the default branch?
6 That question depends on the number of variables used. Since you did not specify what compiler or language or even operating system, that is a difficult one to pin down on! It all rests with the operating system who is responsible for the memory management of the applicat...
Segments of memory - BSS, Stack, Heap, Data, Code/Text (Are there any more?). Say I have a 128MB RAM, Can someone tell me: How much memory is allocated for each of these memory segments? Where do they start? Please specify the address range or something like that for better clarity. What factors influence which shou...
BSS, Stack, Heap, Data, Code/Text - Where each of these start in memory?
The documentation states If you are using Docker Machine on Mac or Windows, your Docker daemon has only limited access to your OS X or Windows filesystem. Docker Machine tries to auto-share your /Users (OS X) or C:\Users (Windows) directory. So, you can mount files or directories on OS X using. docker run -v /U...
I'm trying to run a docker-compose on my Windows machine spinning up a MongoDB instance and a data-only container which proxies an attached volume containing the database files. mongodata: image: mongo:2.6.8 volumes: - ./data/db:/data/db command: --break-mongo mongo: image: mongo:2.6.8 volumes_from: ...
Docker cannot start MongoDb with attached volume through data-only container
Upgrading to Windows 10 Pro and re-installing docker did the trick
I am trying to get an existing docker app running on Windows 10. Problem 1: When I install the docker toolbox, the Docker Quickstart Terminal is missing Problem 2: When I run docker build, I get this error ERROR: Windows named pipe error: The system cannot find the file specified. (code: 2) Any ideas how to solve both...
Docker and Windows 10
Here's a good tutorial on what crontab is and how to use it on Ubuntu. Your crontab line will look something like this: 00 00 * * * ruby path/to/your/script.rb (00 00 indicates midnight--0 minutes and 0 hours--and the *s mean every day of every month.) Syntax: mm hh dd mt wd command mm minute 0-59 hh hour 0...
I have heard crontab is a good choice, but how do I write the line and where do I put it on the server?
How to write a cron that will run a script every day at midnight?
2012-10-17 from the examples is a static, constant, literal date expression, specifying the version of IAM policy language your policy statement uses -- not your policy statement's revision date. There are only two possible values that you can use here, as of now: 2012-10-17 and 2008-10-17. If you don't specify, it's...
I wanted to provide public access to all files in my bucket. Several SO answers including this popular one indicated that I should create a policy. So I went and copy pasted and edited the resource name and version date, but I get an error Document is invalid: Invalid Version 2014-05-02 - undefined I went and looked ...
Amazon S3 bucket policies don't support "version" option