prompt
stringlengths
77
6.9k
completion
stringlengths
1
3.38k
text
stringlengths
163
9.03k
commit
stringlengths
40
40
old_file
stringlengths
4
264
new_file
stringlengths
4
264
lang
stringclasses
277 values
<|file_sep|>original/lib/SimpleSAML/Error/UnserializableException.php <|file_sep|>current/lib/SimpleSAML/Error/UnserializableException.php <|file_sep|>updated/lib/SimpleSAML/Error/UnserializableException.php
<?php /** * Class for saving normal exceptions for serialization. * * This class is used by the SimpleSAML_Auth_State class when it needs * to serialize an exception which doesn't subclass the * SimpleSAML_Error_Exception class. * * It creates a new exception which contains the backtrace and message * of the original exception. * * @package simpleSAMLphp * @version $Id$ */ class SimpleSAML_Error_UnserializableException extends SimpleSAML_Error_Exception { public function __construct(Exception $original) { $msg = get_class($original) . ': ' . $original->getMessage(); $code = $original->getCode();
<|file_sep|>original/lib/SimpleSAML/Error/UnserializableException.php <|file_sep|>current/lib/SimpleSAML/Error/UnserializableException.php <|file_sep|>updated/lib/SimpleSAML/Error/UnserializableException.php <?php /** * Class for saving normal exceptions for serialization. * * This class is used by the SimpleSAML_Auth_State class when it needs * to serialize an exception which doesn't subclass the * SimpleSAML_Error_Exception class. * * It creates a new exception which contains the backtrace and message * of the original exception. * * @package simpleSAMLphp * @version $Id$ */ class SimpleSAML_Error_UnserializableException extends SimpleSAML_Error_Exception { public function __construct(Exception $original) { $msg = get_class($original) . ': ' . $original->getMessage(); $code = $original->getCode();
c2d23ab92293d8059943b5cbaa3b7a2b6d49266b
lib/SimpleSAML/Error/UnserializableException.php
lib/SimpleSAML/Error/UnserializableException.php
PHP
<|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff original: { stdenv, lib, fetchFromGitHub, buildGoModule }: updated: { stdenv, lib, fetchFromGitHub, buildGoModule, fetchpatch }: <|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff original: updated: patches = [ # Fix floating point imprecision issue in the test suite. # https://github.com/google/starlark-go/pull/409 (fetchpatch { url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch"; sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k="; }) ]; <|file_sep|>original/pkgs/development/interpreters/starlark/default.nix version = "unstable-2022-03-02"; src = fetchFromGitHub { owner = "google"; repo = "starlark-go"; rev = "5411bad688d12781515a91cc032645331b4fc302"; sha256 = "sha256-JNsGyGlIVMS5w0W4jHVsrPqqNms3Xfpa4n/XcEWqt6I="; }; vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY="; ldflags = [ "-s" "-w" ]; meta = with lib; { broken = (stdenv.isLinux && stdenv.isAarch64); homepage = "https://github.com/google/starlark-go"; description = "An interpreter for Starlark, implemented in Go"; license = licenses.bsd3; maintainers = with maintainers; [ aaronjheng ]; }; } <|file_sep|>current/pkgs/development/interpreters/starlark/default.nix vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY="; patches = [ # Fix floating point imprecision issue in the test suite. # https://github.com/google/starlark-go/pull/409 (fetchpatch { url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch"; sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k="; }) ]; ldflags = [ "-s" "-w" ]; meta = with lib; { broken = (stdenv.isLinux && stdenv.isAarch64); homepage = "https://github.com/google/starlark-go"; description = "An interpreter for Starlark, implemented in Go"; license = licenses.bsd3; maintainers = with maintainers; [ aaronjheng ]; }; } <|file_sep|>updated/pkgs/development/interpreters/starlark/default.nix
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY="; patches = [ # Fix floating point imprecision issue in the test suite. # https://github.com/google/starlark-go/pull/409 (fetchpatch { url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch"; sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k="; }) ]; ldflags = [ "-s" "-w" ]; meta = with lib; { homepage = "https://github.com/google/starlark-go"; description = "An interpreter for Starlark, implemented in Go"; license = licenses.bsd3; maintainers = with maintainers; [ aaronjheng ]; }; }
<|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff original: { stdenv, lib, fetchFromGitHub, buildGoModule }: updated: { stdenv, lib, fetchFromGitHub, buildGoModule, fetchpatch }: <|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff original: updated: patches = [ # Fix floating point imprecision issue in the test suite. # https://github.com/google/starlark-go/pull/409 (fetchpatch { url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch"; sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k="; }) ]; <|file_sep|>original/pkgs/development/interpreters/starlark/default.nix version = "unstable-2022-03-02"; src = fetchFromGitHub { owner = "google"; repo = "starlark-go"; rev = "5411bad688d12781515a91cc032645331b4fc302"; sha256 = "sha256-JNsGyGlIVMS5w0W4jHVsrPqqNms3Xfpa4n/XcEWqt6I="; }; vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY="; ldflags = [ "-s" "-w" ]; meta = with lib; { broken = (stdenv.isLinux && stdenv.isAarch64); homepage = "https://github.com/google/starlark-go"; description = "An interpreter for Starlark, implemented in Go"; license = licenses.bsd3; maintainers = with maintainers; [ aaronjheng ]; }; } <|file_sep|>current/pkgs/development/interpreters/starlark/default.nix vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY="; patches = [ # Fix floating point imprecision issue in the test suite. # https://github.com/google/starlark-go/pull/409 (fetchpatch { url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch"; sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k="; }) ]; ldflags = [ "-s" "-w" ]; meta = with lib; { broken = (stdenv.isLinux && stdenv.isAarch64); homepage = "https://github.com/google/starlark-go"; description = "An interpreter for Starlark, implemented in Go"; license = licenses.bsd3; maintainers = with maintainers; [ aaronjheng ]; }; } <|file_sep|>updated/pkgs/development/interpreters/starlark/default.nix vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY="; patches = [ # Fix floating point imprecision issue in the test suite. # https://github.com/google/starlark-go/pull/409 (fetchpatch { url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch"; sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k="; }) ]; ldflags = [ "-s" "-w" ]; meta = with lib; { homepage = "https://github.com/google/starlark-go"; description = "An interpreter for Starlark, implemented in Go"; license = licenses.bsd3; maintainers = with maintainers; [ aaronjheng ]; }; }
699a374e8a24edbd3296240f6da5d454e7fb3c21
pkgs/development/interpreters/starlark/default.nix
pkgs/development/interpreters/starlark/default.nix
Nix
<|file_sep|>original/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy } } catch (IOException ignored) { // ignore } } /** * Create a progress logger for an arbitrary project and class. * * @param project the project to create a ProgressLogger for. * @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null. * @return instance of ProgressLogger. */ static ProgressLogger getProgressLogger(final Project project, final Class clazz) { ServiceRegistry registry = (project.gradle as GradleInternal).getServices() ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory) ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz)) progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}") progressLogger.setLoggingHeader(null) } } <|file_sep|>current/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy } } catch (IOException ignored) { // ignore } } /** * Create a progress logger for an arbitrary project and class. * * @param project the project to create a ProgressLogger for. * @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null. * @return instance of ProgressLogger. */ static ProgressLogger getProgressLogger(final Project project, final Class clazz) { ServiceRegistry registry = (project.gradle as GradleInternal).getServices() ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory) ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz)) progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}") progressLogger.setLoggingHeader(null) } } <|file_sep|>updated/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
toClose.close() } } catch (IOException ignored) { // ignore } } /** * Create a progress logger for an arbitrary project and class. * * @param project the project to create a ProgressLogger for. * @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null. * @return instance of ProgressLogger. */ static ProgressLogger getProgressLogger(final Project project, final Class clazz) { ServiceRegistry registry = (project.gradle as GradleInternal).getServices() ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory) ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz)) progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}") } }
<|file_sep|>original/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy } } catch (IOException ignored) { // ignore } } /** * Create a progress logger for an arbitrary project and class. * * @param project the project to create a ProgressLogger for. * @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null. * @return instance of ProgressLogger. */ static ProgressLogger getProgressLogger(final Project project, final Class clazz) { ServiceRegistry registry = (project.gradle as GradleInternal).getServices() ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory) ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz)) progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}") progressLogger.setLoggingHeader(null) } } <|file_sep|>current/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy } } catch (IOException ignored) { // ignore } } /** * Create a progress logger for an arbitrary project and class. * * @param project the project to create a ProgressLogger for. * @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null. * @return instance of ProgressLogger. */ static ProgressLogger getProgressLogger(final Project project, final Class clazz) { ServiceRegistry registry = (project.gradle as GradleInternal).getServices() ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory) ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz)) progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}") progressLogger.setLoggingHeader(null) } } <|file_sep|>updated/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy toClose.close() } } catch (IOException ignored) { // ignore } } /** * Create a progress logger for an arbitrary project and class. * * @param project the project to create a ProgressLogger for. * @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null. * @return instance of ProgressLogger. */ static ProgressLogger getProgressLogger(final Project project, final Class clazz) { ServiceRegistry registry = (project.gradle as GradleInternal).getServices() ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory) ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz)) progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}") } }
c0ae71b52281cdd107aacf56286af97d2b4b9d6b
src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
Groovy
<|file_sep|>original/app.json "description":"Send and receive messages from a Habitica Party Chat to a Slack Channel", "repository":"https://github.com/niteshpatel/habitica-slack", "keywords":[ "habitica", "slack" ], "env":{ "DJANGO_SECRETKEY":{ "description":"Secret key for your Django instance (autogenerated).", "generator":"secret" }, "HABITICA_APIUSER":{ "description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_APIKEY":{ "description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_GROUPID":{ "description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)." }, "SLACK_TOKEN":{ <|file_sep|>current/app.json "description":"Send and receive messages from a Habitica Party Chat to a Slack Channel", "repository":"https://github.com/niteshpatel/habitica-slack", "keywords":[ "habitica", "slack" ], "env":{ "DJANGO_SECRETKEY":{ "description":"Secret key for your Django instance (autogenerated).", "generator":"secret" }, "HABITICA_APIUSER":{ "description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_APIKEY":{ "description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_GROUPID":{ "description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)." }, "SLACK_TOKEN":{ <|file_sep|>updated/app.json
"description": "Send and receive messages from a Habitica Party Chat to a Slack Channel", "repository": "https://github.com/niteshpatel/habitica-slack", "keywords": [ "habitica", "slack" ], "env": { "DJANGO_SECRETKEY": { "description": "Secret key for your Django instance (autogenerated).", "generator": "secret" }, "HABITICA_APIUSER": { "description": "Your Habitica User ID (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_APIKEY": { "description": "Your Habitica API Token (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_GROUPID": { "description": "Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)." }, "SLACK_TOKEN": {
<|file_sep|>original/app.json "description":"Send and receive messages from a Habitica Party Chat to a Slack Channel", "repository":"https://github.com/niteshpatel/habitica-slack", "keywords":[ "habitica", "slack" ], "env":{ "DJANGO_SECRETKEY":{ "description":"Secret key for your Django instance (autogenerated).", "generator":"secret" }, "HABITICA_APIUSER":{ "description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_APIKEY":{ "description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_GROUPID":{ "description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)." }, "SLACK_TOKEN":{ <|file_sep|>current/app.json "description":"Send and receive messages from a Habitica Party Chat to a Slack Channel", "repository":"https://github.com/niteshpatel/habitica-slack", "keywords":[ "habitica", "slack" ], "env":{ "DJANGO_SECRETKEY":{ "description":"Secret key for your Django instance (autogenerated).", "generator":"secret" }, "HABITICA_APIUSER":{ "description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_APIKEY":{ "description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_GROUPID":{ "description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)." }, "SLACK_TOKEN":{ <|file_sep|>updated/app.json "description": "Send and receive messages from a Habitica Party Chat to a Slack Channel", "repository": "https://github.com/niteshpatel/habitica-slack", "keywords": [ "habitica", "slack" ], "env": { "DJANGO_SECRETKEY": { "description": "Secret key for your Django instance (autogenerated).", "generator": "secret" }, "HABITICA_APIUSER": { "description": "Your Habitica User ID (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_APIKEY": { "description": "Your Habitica API Token (found here https://habitica.com/#/options/settings/api)." }, "HABITICA_GROUPID": { "description": "Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)." }, "SLACK_TOKEN": {
9728be88389c1ff45fd79413bb3751f2b74e9c50
app.json
app.json
JSON
<|file_sep|>original/admin/app/views/transactions/forms/_delete.html.erb <%= form_tag("/generate_delete_transaction", remote: true) do %> <%= text_field_tag 'public_key', creator_address, hidden: true %> <%= text_field_tag 'payload', payload, hidden: true %> <%= text_field_tag 'priv_key', private_key, hidden: true %> <%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %> <% end %> <|file_sep|>current/admin/app/views/transactions/forms/_delete.html.erb <%= form_tag("/generate_delete_transaction", remote: true) do %> <%= text_field_tag 'public_key', creator_address, hidden: true %> <%= text_field_tag 'payload', payload, hidden: true %> <%= text_field_tag 'priv_key', private_key, hidden: true %> <%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %> <% end %> <|file_sep|>updated/admin/app/views/transactions/forms/_delete.html.erb
<%= form_tag("/generate_delete_transaction", remote: true) do %> <%= text_field_tag 'public_key', creator_address, hidden: true %> <%= text_field_tag 'payload', payload, hidden: true %> <%= text_field_tag 'priv_key', private_key, hidden: true %> <%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Accept') ? 'Accepting...' : 'Cancelling...'}" } %> <% end %>
<|file_sep|>original/admin/app/views/transactions/forms/_delete.html.erb <%= form_tag("/generate_delete_transaction", remote: true) do %> <%= text_field_tag 'public_key', creator_address, hidden: true %> <%= text_field_tag 'payload', payload, hidden: true %> <%= text_field_tag 'priv_key', private_key, hidden: true %> <%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %> <% end %> <|file_sep|>current/admin/app/views/transactions/forms/_delete.html.erb <%= form_tag("/generate_delete_transaction", remote: true) do %> <%= text_field_tag 'public_key', creator_address, hidden: true %> <%= text_field_tag 'payload', payload, hidden: true %> <%= text_field_tag 'priv_key', private_key, hidden: true %> <%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %> <% end %> <|file_sep|>updated/admin/app/views/transactions/forms/_delete.html.erb <%= form_tag("/generate_delete_transaction", remote: true) do %> <%= text_field_tag 'public_key', creator_address, hidden: true %> <%= text_field_tag 'payload', payload, hidden: true %> <%= text_field_tag 'priv_key', private_key, hidden: true %> <%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Accept') ? 'Accepting...' : 'Cancelling...'}" } %> <% end %>
fe87adc4d4567f2c162a5bffb97f21c7d819550c
admin/app/views/transactions/forms/_delete.html.erb
admin/app/views/transactions/forms/_delete.html.erb
HTML+ERB
<|file_sep|>original/common_audio/signal_processing/cross_correlation.c /* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */ void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation, const int16_t* seq1, const int16_t* seq2, int16_t dim_seq, int16_t dim_cross_correlation, int right_shifts, int step_seq2) { int i = 0, j = 0; for (i = 0; i < dim_cross_correlation; i++) { int32_t corr = 0; /* Unrolling doesn't seem to improve performance. */ for (j = 0; j < dim_seq; j++) { // It's not clear why casting |right_shifts| here helps performance. corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts; } seq2 += step_seq2; *cross_correlation++ = corr; } } <|file_sep|>current/common_audio/signal_processing/cross_correlation.c /* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */ void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation, const int16_t* seq1, const int16_t* seq2, int16_t dim_seq, int16_t dim_cross_correlation, int right_shifts, int step_seq2) { int i = 0, j = 0; for (i = 0; i < dim_cross_correlation; i++) { int32_t corr = 0; /* Unrolling doesn't seem to improve performance. */ for (j = 0; j < dim_seq; j++) { // It's not clear why casting |right_shifts| here helps performance. corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts; } seq2 += step_seq2; *cross_correlation++ = corr; } } <|file_sep|>updated/common_audio/signal_processing/cross_correlation.c
#include "webrtc/common_audio/signal_processing/include/signal_processing_library.h" /* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */ void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation, const int16_t* seq1, const int16_t* seq2, int16_t dim_seq, int16_t dim_cross_correlation, int right_shifts, int step_seq2) { int i = 0, j = 0; for (i = 0; i < dim_cross_correlation; i++) { int32_t corr = 0; /* Unrolling doesn't seem to improve performance. */ for (j = 0; j < dim_seq; j++) corr += (seq1[j] * seq2[j]) >> right_shifts; seq2 += step_seq2; *cross_correlation++ = corr; } }
<|file_sep|>original/common_audio/signal_processing/cross_correlation.c /* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */ void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation, const int16_t* seq1, const int16_t* seq2, int16_t dim_seq, int16_t dim_cross_correlation, int right_shifts, int step_seq2) { int i = 0, j = 0; for (i = 0; i < dim_cross_correlation; i++) { int32_t corr = 0; /* Unrolling doesn't seem to improve performance. */ for (j = 0; j < dim_seq; j++) { // It's not clear why casting |right_shifts| here helps performance. corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts; } seq2 += step_seq2; *cross_correlation++ = corr; } } <|file_sep|>current/common_audio/signal_processing/cross_correlation.c /* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */ void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation, const int16_t* seq1, const int16_t* seq2, int16_t dim_seq, int16_t dim_cross_correlation, int right_shifts, int step_seq2) { int i = 0, j = 0; for (i = 0; i < dim_cross_correlation; i++) { int32_t corr = 0; /* Unrolling doesn't seem to improve performance. */ for (j = 0; j < dim_seq; j++) { // It's not clear why casting |right_shifts| here helps performance. corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts; } seq2 += step_seq2; *cross_correlation++ = corr; } } <|file_sep|>updated/common_audio/signal_processing/cross_correlation.c #include "webrtc/common_audio/signal_processing/include/signal_processing_library.h" /* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */ void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation, const int16_t* seq1, const int16_t* seq2, int16_t dim_seq, int16_t dim_cross_correlation, int right_shifts, int step_seq2) { int i = 0, j = 0; for (i = 0; i < dim_cross_correlation; i++) { int32_t corr = 0; /* Unrolling doesn't seem to improve performance. */ for (j = 0; j < dim_seq; j++) corr += (seq1[j] * seq2[j]) >> right_shifts; seq2 += step_seq2; *cross_correlation++ = corr; } }
8504c7c838219029c7e92818f663c777b8bbe5b6
common_audio/signal_processing/cross_correlation.c
common_audio/signal_processing/cross_correlation.c
C
<|file_sep|>original/roles/simple/graphviz/tasks/main.yml - name: Install graphviz homebrew: name: graphviz install_options: with-app state: present when: ansible_distribution == "MacOSX" - name: Install graphviz become: yes apt: name: graphviz state: present when: ansible_os_family == "Debian" - name: Install graphviz become: yes pkgng: name: graphviz state: present when: ansible_distribution == "FreeBSD" <|file_sep|>current/roles/simple/graphviz/tasks/main.yml - name: Install graphviz homebrew: name: graphviz install_options: with-app state: present when: ansible_distribution == "MacOSX" - name: Install graphviz become: yes apt: name: graphviz state: present when: ansible_os_family == "Debian" - name: Install graphviz become: yes pkgng: name: graphviz state: present when: ansible_distribution == "FreeBSD" <|file_sep|>updated/roles/simple/graphviz/tasks/main.yml
- name: Install graphviz homebrew: name: graphviz install_options: with-app, with-bindings state: present when: ansible_distribution == "MacOSX" - name: Install graphviz become: yes apt: name: graphviz state: present when: ansible_os_family == "Debian" - name: Install graphviz become: yes pkgng: name: graphviz state: present when: ansible_distribution == "FreeBSD"
<|file_sep|>original/roles/simple/graphviz/tasks/main.yml - name: Install graphviz homebrew: name: graphviz install_options: with-app state: present when: ansible_distribution == "MacOSX" - name: Install graphviz become: yes apt: name: graphviz state: present when: ansible_os_family == "Debian" - name: Install graphviz become: yes pkgng: name: graphviz state: present when: ansible_distribution == "FreeBSD" <|file_sep|>current/roles/simple/graphviz/tasks/main.yml - name: Install graphviz homebrew: name: graphviz install_options: with-app state: present when: ansible_distribution == "MacOSX" - name: Install graphviz become: yes apt: name: graphviz state: present when: ansible_os_family == "Debian" - name: Install graphviz become: yes pkgng: name: graphviz state: present when: ansible_distribution == "FreeBSD" <|file_sep|>updated/roles/simple/graphviz/tasks/main.yml - name: Install graphviz homebrew: name: graphviz install_options: with-app, with-bindings state: present when: ansible_distribution == "MacOSX" - name: Install graphviz become: yes apt: name: graphviz state: present when: ansible_os_family == "Debian" - name: Install graphviz become: yes pkgng: name: graphviz state: present when: ansible_distribution == "FreeBSD"
d9f054c1a36da903820e6ebbd444ad926408fb8c
roles/simple/graphviz/tasks/main.yml
roles/simple/graphviz/tasks/main.yml
YAML
<|file_sep|>tests/parser/parserTests.txt.diff original: !! Version 2 updated: !! options version=2 parsoid-compatible requirements=extension:Disambiguator !! end <|file_sep|>original/tests/parser/parserTests.txt !! article Disambig !! text This is a disambig page. __DISAMBIG__ !! endarticle !! article Normal !! text This is a normal page. !! endarticle !! test Link to disambig and non-disambig page !! wikitext I am linking to [[Disambig]] and [[Normal]]. !! html <p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>. </p> !! end <|file_sep|>current/tests/parser/parserTests.txt !! article Disambig !! text This is a disambig page. __DISAMBIG__ !! endarticle !! article Normal !! text This is a normal page. !! endarticle !! test Link to disambig and non-disambig page !! wikitext I am linking to [[Disambig]] and [[Normal]]. !! html <p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>. </p> !! end <|file_sep|>updated/tests/parser/parserTests.txt
This is a normal page. !! endarticle !! test Link to disambig and non-disambig page !! wikitext I am linking to [[Disambig]] and [[Normal]]. !! html <p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>. </p> !! end !! test Ensure disambiguation links are marked properly in Parsoid !! options parsoid=wt2html !! wikitext [[Disambig]] !! html/parsoid <p><a rel="mw:WikiLink" href="./Disambig" title="Disambig" class="mw-disambig">Disambig</a></p> !! end
<|file_sep|>tests/parser/parserTests.txt.diff original: !! Version 2 updated: !! options version=2 parsoid-compatible requirements=extension:Disambiguator !! end <|file_sep|>original/tests/parser/parserTests.txt !! article Disambig !! text This is a disambig page. __DISAMBIG__ !! endarticle !! article Normal !! text This is a normal page. !! endarticle !! test Link to disambig and non-disambig page !! wikitext I am linking to [[Disambig]] and [[Normal]]. !! html <p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>. </p> !! end <|file_sep|>current/tests/parser/parserTests.txt !! article Disambig !! text This is a disambig page. __DISAMBIG__ !! endarticle !! article Normal !! text This is a normal page. !! endarticle !! test Link to disambig and non-disambig page !! wikitext I am linking to [[Disambig]] and [[Normal]]. !! html <p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>. </p> !! end <|file_sep|>updated/tests/parser/parserTests.txt This is a normal page. !! endarticle !! test Link to disambig and non-disambig page !! wikitext I am linking to [[Disambig]] and [[Normal]]. !! html <p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>. </p> !! end !! test Ensure disambiguation links are marked properly in Parsoid !! options parsoid=wt2html !! wikitext [[Disambig]] !! html/parsoid <p><a rel="mw:WikiLink" href="./Disambig" title="Disambig" class="mw-disambig">Disambig</a></p> !! end
f9e5a56bf8d41d90b3dd220e7c8497a56c8f62ce
tests/parser/parserTests.txt
tests/parser/parserTests.txt
Text
<|file_sep|>original/assets/be/views/elements/dataTableForm.html <div class="detail" ng-controller="FormEditController as fec"> <form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'> <formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form"> <div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)"> <!-- button to show form --> <button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button> <!-- buttons to submit / cancel form --> <span ng-show="fec.form.$visible"> <button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button> <button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button> </span> </div> </formly-form> </form> </div> <|file_sep|>current/assets/be/views/elements/dataTableForm.html <div class="detail" ng-controller="FormEditController as fec"> <form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'> <formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form"> <div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)"> <!-- button to show form --> <button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button> <!-- buttons to submit / cancel form --> <span ng-show="fec.form.$visible"> <button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button> <button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button> </span> </div> </formly-form> </form> </div> <|file_sep|>updated/assets/be/views/elements/dataTableForm.html
<div class="detail" ng-controller="FormEditController as fec"> <form novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'> <formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form"> <div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)"> <!-- button to show form --> <button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button> <!-- buttons to submit / cancel form --> <span ng-show="fec.form.$visible"> <button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button> <button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button> </span> </div> </formly-form> </form> </div>
<|file_sep|>original/assets/be/views/elements/dataTableForm.html <div class="detail" ng-controller="FormEditController as fec"> <form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'> <formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form"> <div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)"> <!-- button to show form --> <button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button> <!-- buttons to submit / cancel form --> <span ng-show="fec.form.$visible"> <button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button> <button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button> </span> </div> </formly-form> </form> </div> <|file_sep|>current/assets/be/views/elements/dataTableForm.html <div class="detail" ng-controller="FormEditController as fec"> <form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'> <formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form"> <div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)"> <!-- button to show form --> <button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button> <!-- buttons to submit / cancel form --> <span ng-show="fec.form.$visible"> <button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button> <button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button> </span> </div> </formly-form> </form> </div> <|file_sep|>updated/assets/be/views/elements/dataTableForm.html <div class="detail" ng-controller="FormEditController as fec"> <form novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'> <formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form"> <div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)"> <!-- button to show form --> <button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button> <!-- buttons to submit / cancel form --> <span ng-show="fec.form.$visible"> <button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button> <button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button> </span> </div> </formly-form> </form> </div>
de4202964111fc6aa792122052c19a78d8902e6b
assets/be/views/elements/dataTableForm.html
assets/be/views/elements/dataTableForm.html
HTML
<|file_sep|>original/bower.json "dist/neue.css", "dist/neue.min.js" ], "dependencies": { "jquery": "^1.7", "susy": "2.2.0.rc.2" }, "devDependencies": { "jquery": "1.8.3", "almond": "~0.2.9", "html5shiv": "~3.7.2", "respond": "~1.4.2", "highlightjs": "~8.0.0", "filament-sticky": "~0.1.4", "dosomething-modal": "DoSomething/modal#~0.1.0", "dosomething-validation": "DoSomething/validation#~0.0.3" }, "resolutions": { "jquery": "1.8.3" } } <|file_sep|>current/bower.json "dist/neue.css", "dist/neue.min.js" ], "dependencies": { "jquery": "^1.7", "susy": "2.2.0.rc.2" }, "devDependencies": { "jquery": "1.8.3", "almond": "~0.2.9", "html5shiv": "~3.7.2", "respond": "~1.4.2", "highlightjs": "~8.0.0", "filament-sticky": "~0.1.4", "dosomething-modal": "DoSomething/modal#~0.1.0", "dosomething-validation": "DoSomething/validation#~0.0.3" }, "resolutions": { "jquery": "1.8.3" } } <|file_sep|>updated/bower.json
"dist/neue.css", "dist/neue.min.js" ], "dependencies": { "jquery": "^1.7", "susy": "2.2.0.rc.2" }, "devDependencies": { "jquery": "1.8.3", "almond": "~0.2.9", "html5shiv": "~3.7.2", "respond": "~1.4.2", "highlightjs": "~8.0.0", "filament-sticky": "~0.1.4", "dosomething-modal": "DoSomething/modal#~0.1.0", "dosomething-validation": "DoSomething/validation#~0.1.0" }, "resolutions": { "jquery": "1.8.3" } }
<|file_sep|>original/bower.json "dist/neue.css", "dist/neue.min.js" ], "dependencies": { "jquery": "^1.7", "susy": "2.2.0.rc.2" }, "devDependencies": { "jquery": "1.8.3", "almond": "~0.2.9", "html5shiv": "~3.7.2", "respond": "~1.4.2", "highlightjs": "~8.0.0", "filament-sticky": "~0.1.4", "dosomething-modal": "DoSomething/modal#~0.1.0", "dosomething-validation": "DoSomething/validation#~0.0.3" }, "resolutions": { "jquery": "1.8.3" } } <|file_sep|>current/bower.json "dist/neue.css", "dist/neue.min.js" ], "dependencies": { "jquery": "^1.7", "susy": "2.2.0.rc.2" }, "devDependencies": { "jquery": "1.8.3", "almond": "~0.2.9", "html5shiv": "~3.7.2", "respond": "~1.4.2", "highlightjs": "~8.0.0", "filament-sticky": "~0.1.4", "dosomething-modal": "DoSomething/modal#~0.1.0", "dosomething-validation": "DoSomething/validation#~0.0.3" }, "resolutions": { "jquery": "1.8.3" } } <|file_sep|>updated/bower.json "dist/neue.css", "dist/neue.min.js" ], "dependencies": { "jquery": "^1.7", "susy": "2.2.0.rc.2" }, "devDependencies": { "jquery": "1.8.3", "almond": "~0.2.9", "html5shiv": "~3.7.2", "respond": "~1.4.2", "highlightjs": "~8.0.0", "filament-sticky": "~0.1.4", "dosomething-modal": "DoSomething/modal#~0.1.0", "dosomething-validation": "DoSomething/validation#~0.1.0" }, "resolutions": { "jquery": "1.8.3" } }
08f1a3d5d15ab79c6c0ce7467e570acbee30b568
bower.json
bower.json
JSON
<|file_sep|>Tools/idle/ZoomHeight.py.diff original: height = top.winfo_screenheight() updated: newheight = top.winfo_screenheight() <|file_sep|>Tools/idle/ZoomHeight.py.diff original: y = 0 height = height - 72 updated: newy = 0 newheight = newheight - 72 <|file_sep|>Tools/idle/ZoomHeight.py.diff original: y = 24 height = height - 96 newgeom = "%dx%d+%d+%d" % (width, height, x, y) if geom == newgeom: updated: newy = 24 newheight = newheight - 96 if height >= newheight: <|file_sep|>original/Tools/idle/ZoomHeight.py def __init__(self, editwin): self.editwin = editwin def zoom_height_event(self, event): top = self.editwin.top geom = top.wm_geometry() m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom) if not m: top.bell() return width, height, x, y = map(int, m.groups()) height = top.winfo_screenheight() if sys.platform == 'win32': y = 0 height = height - 72 else: y = 24 height = height - 96 newgeom = "%dx%d+%d+%d" % (width, height, x, y) if geom == newgeom: newgeom = "" <|file_sep|>current/Tools/idle/ZoomHeight.py def __init__(self, editwin): self.editwin = editwin def zoom_height_event(self, event): top = self.editwin.top geom = top.wm_geometry() m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom) if not m: top.bell() return width, height, x, y = map(int, m.groups()) newheight = top.winfo_screenheight() if sys.platform == 'win32': newy = 0 newheight = newheight - 72 else: newy = 24 newheight = newheight - 96 if height >= newheight: newgeom = "" top.wm_geometry(newgeom) <|file_sep|>updated/Tools/idle/ZoomHeight.py
def zoom_height_event(self, event): top = self.editwin.top geom = top.wm_geometry() m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom) if not m: top.bell() return width, height, x, y = map(int, m.groups()) newheight = top.winfo_screenheight() if sys.platform == 'win32': newy = 0 newheight = newheight - 72 else: newy = 24 newheight = newheight - 96 if height >= newheight: newgeom = "" else: newgeom = "%dx%d+%d+%d" % (width, newheight, x, newy) top.wm_geometry(newgeom)
<|file_sep|>Tools/idle/ZoomHeight.py.diff original: height = top.winfo_screenheight() updated: newheight = top.winfo_screenheight() <|file_sep|>Tools/idle/ZoomHeight.py.diff original: y = 0 height = height - 72 updated: newy = 0 newheight = newheight - 72 <|file_sep|>Tools/idle/ZoomHeight.py.diff original: y = 24 height = height - 96 newgeom = "%dx%d+%d+%d" % (width, height, x, y) if geom == newgeom: updated: newy = 24 newheight = newheight - 96 if height >= newheight: <|file_sep|>original/Tools/idle/ZoomHeight.py def __init__(self, editwin): self.editwin = editwin def zoom_height_event(self, event): top = self.editwin.top geom = top.wm_geometry() m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom) if not m: top.bell() return width, height, x, y = map(int, m.groups()) height = top.winfo_screenheight() if sys.platform == 'win32': y = 0 height = height - 72 else: y = 24 height = height - 96 newgeom = "%dx%d+%d+%d" % (width, height, x, y) if geom == newgeom: newgeom = "" <|file_sep|>current/Tools/idle/ZoomHeight.py def __init__(self, editwin): self.editwin = editwin def zoom_height_event(self, event): top = self.editwin.top geom = top.wm_geometry() m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom) if not m: top.bell() return width, height, x, y = map(int, m.groups()) newheight = top.winfo_screenheight() if sys.platform == 'win32': newy = 0 newheight = newheight - 72 else: newy = 24 newheight = newheight - 96 if height >= newheight: newgeom = "" top.wm_geometry(newgeom) <|file_sep|>updated/Tools/idle/ZoomHeight.py def zoom_height_event(self, event): top = self.editwin.top geom = top.wm_geometry() m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom) if not m: top.bell() return width, height, x, y = map(int, m.groups()) newheight = top.winfo_screenheight() if sys.platform == 'win32': newy = 0 newheight = newheight - 72 else: newy = 24 newheight = newheight - 96 if height >= newheight: newgeom = "" else: newgeom = "%dx%d+%d+%d" % (width, newheight, x, newy) top.wm_geometry(newgeom)
ffd1ba9eee804fdd55e86908b158a4ad94f2f366
Tools/idle/ZoomHeight.py
Tools/idle/ZoomHeight.py
Python
<|file_sep|>original/config/fish/env.fish append-to-path ~/projects/git-toolbelt set -gx WORKON_HOME ~/.virtualenvs set -gx PYTHONPATH ~/.local/bin append-to-path ~/.local/bin append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin <|file_sep|>current/config/fish/env.fish append-to-path ~/projects/git-toolbelt set -gx WORKON_HOME ~/.virtualenvs set -gx PYTHONPATH ~/.local/bin append-to-path ~/.local/bin append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin <|file_sep|>updated/config/fish/env.fish
append-to-path ~/projects/git-toolbelt set -gx WORKON_HOME ~/.virtualenvs set -gx PYTHONPATH ~/.local/bin append-to-path ~/.local/bin append-to-path /Applications/Postgres.app/Contents/Versions/9.6/bin
<|file_sep|>original/config/fish/env.fish append-to-path ~/projects/git-toolbelt set -gx WORKON_HOME ~/.virtualenvs set -gx PYTHONPATH ~/.local/bin append-to-path ~/.local/bin append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin <|file_sep|>current/config/fish/env.fish append-to-path ~/projects/git-toolbelt set -gx WORKON_HOME ~/.virtualenvs set -gx PYTHONPATH ~/.local/bin append-to-path ~/.local/bin append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin <|file_sep|>updated/config/fish/env.fish append-to-path ~/projects/git-toolbelt set -gx WORKON_HOME ~/.virtualenvs set -gx PYTHONPATH ~/.local/bin append-to-path ~/.local/bin append-to-path /Applications/Postgres.app/Contents/Versions/9.6/bin
af1f95bde61eb9553d26ca6ee3adf8e95ca8e35b
config/fish/env.fish
config/fish/env.fish
fish
<|file_sep|>index.js.diff original: updated: var confluenceUrl = process.env.CONFLUENCE_URL <|file_sep|>original/index.js #!/usr/bin/env node var program = require('commander'); var userArgs = process.argv.splice(2); var message = userArgs.join(' '); if (message.length > 140) { console.log('Message was too long. Can only be 140 characters. It was: ', message.length); process.exit(1); } console.log(message); <|file_sep|>current/index.js #!/usr/bin/env node var program = require('commander'); var userArgs = process.argv.splice(2); var message = userArgs.join(' '); var confluenceUrl = process.env.CONFLUENCE_URL if (message.length > 140) { console.log('Message was too long. Can only be 140 characters. It was: ', message.length); process.exit(1); } console.log(message); <|file_sep|>updated/index.js
#!/usr/bin/env node var program = require('commander'); var userArgs = process.argv.splice(2); var message = userArgs.join(' '); var confluenceUrl = process.env.CONFLUENCE_URL if (message.length > 140) { console.log('Message was too long. Can only be 140 characters. It was: ', message.length); process.exit(1); } if (!confluenceUrl) { console.log('Please set the environment variable CONFLUENCE_URL.') process.exit(2); } console.log(message);
<|file_sep|>index.js.diff original: updated: var confluenceUrl = process.env.CONFLUENCE_URL <|file_sep|>original/index.js #!/usr/bin/env node var program = require('commander'); var userArgs = process.argv.splice(2); var message = userArgs.join(' '); if (message.length > 140) { console.log('Message was too long. Can only be 140 characters. It was: ', message.length); process.exit(1); } console.log(message); <|file_sep|>current/index.js #!/usr/bin/env node var program = require('commander'); var userArgs = process.argv.splice(2); var message = userArgs.join(' '); var confluenceUrl = process.env.CONFLUENCE_URL if (message.length > 140) { console.log('Message was too long. Can only be 140 characters. It was: ', message.length); process.exit(1); } console.log(message); <|file_sep|>updated/index.js #!/usr/bin/env node var program = require('commander'); var userArgs = process.argv.splice(2); var message = userArgs.join(' '); var confluenceUrl = process.env.CONFLUENCE_URL if (message.length > 140) { console.log('Message was too long. Can only be 140 characters. It was: ', message.length); process.exit(1); } if (!confluenceUrl) { console.log('Please set the environment variable CONFLUENCE_URL.') process.exit(2); } console.log(message);
65bb0ff4fce04e58febe70936adf503d432f9d41
index.js
index.js
JavaScript
<|file_sep|>original/server/websocket.js var clients = {} wss.on('connection', (ws) => { ws.on('message', (data) => { var req = JSON.parse(data) if (req.message === 'ping') { ws.id = req.id clients[ws.id] = ws clients[ws.id].send(JSON.stringify({message: 'pong'})) } else if (req.message === 'gems') { ds.run(req.gem, (gem) => { if (!clients[req.id]) { return true } clients[req.id].send(JSON.stringify({message: 'node', gem: gem})) }, (err, gems) => { <|file_sep|>current/server/websocket.js var clients = {} wss.on('connection', (ws) => { ws.on('message', (data) => { var req = JSON.parse(data) if (req.message === 'ping') { ws.id = req.id clients[ws.id] = ws clients[ws.id].send(JSON.stringify({message: 'pong'})) } else if (req.message === 'gems') { ds.run(req.gem, (gem) => { if (!clients[req.id]) { return true } clients[req.id].send(JSON.stringify({message: 'node', gem: gem})) }, (err, gems) => { <|file_sep|>updated/server/websocket.js
var clients = {} wss.on('connection', (ws) => { ws.on('message', (data) => { var req = JSON.parse(data) if (req.message === 'ping') { ws.id = req.id clients[ws.id] = ws clients[ws.id].send(JSON.stringify({message: 'pong', id: ws.id})) } else if (req.message === 'gems') { ds.run(req.gem, (gem) => { if (!clients[req.id]) { return true } clients[req.id].send(JSON.stringify({message: 'node', gem: gem})) }, (err, gems) => {
<|file_sep|>original/server/websocket.js var clients = {} wss.on('connection', (ws) => { ws.on('message', (data) => { var req = JSON.parse(data) if (req.message === 'ping') { ws.id = req.id clients[ws.id] = ws clients[ws.id].send(JSON.stringify({message: 'pong'})) } else if (req.message === 'gems') { ds.run(req.gem, (gem) => { if (!clients[req.id]) { return true } clients[req.id].send(JSON.stringify({message: 'node', gem: gem})) }, (err, gems) => { <|file_sep|>current/server/websocket.js var clients = {} wss.on('connection', (ws) => { ws.on('message', (data) => { var req = JSON.parse(data) if (req.message === 'ping') { ws.id = req.id clients[ws.id] = ws clients[ws.id].send(JSON.stringify({message: 'pong'})) } else if (req.message === 'gems') { ds.run(req.gem, (gem) => { if (!clients[req.id]) { return true } clients[req.id].send(JSON.stringify({message: 'node', gem: gem})) }, (err, gems) => { <|file_sep|>updated/server/websocket.js var clients = {} wss.on('connection', (ws) => { ws.on('message', (data) => { var req = JSON.parse(data) if (req.message === 'ping') { ws.id = req.id clients[ws.id] = ws clients[ws.id].send(JSON.stringify({message: 'pong', id: ws.id})) } else if (req.message === 'gems') { ds.run(req.gem, (gem) => { if (!clients[req.id]) { return true } clients[req.id].send(JSON.stringify({message: 'node', gem: gem})) }, (err, gems) => {
af6314d64914d4450ceabb197b2c451e97ac2a62
server/websocket.js
server/websocket.js
JavaScript
<|file_sep|>oh-my-zsh.sh.diff original: # Load all of your custom configurations from custom/ for config_file ($ZSH/custom/*.zsh) source $config_file updated: <|file_sep|>original/oh-my-zsh.sh # Load and run compinit autoload -U compinit compinit -i # Load all of your custom configurations from custom/ for config_file ($ZSH/custom/*.zsh) source $config_file # Load all of the plugins that were defined in ~/.zshrc for plugin ($plugins); do if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then source $ZSH/plugins/$plugin/$plugin.plugin.zsh fi done # Load the theme source "$ZSH/themes/$ZSH_THEME.zsh-theme" # Check for updates on initial load... if [ "$DISABLE_AUTO_UPDATE" = "true" ] then return <|file_sep|>current/oh-my-zsh.sh # Load and run compinit autoload -U compinit compinit -i # Load all of the plugins that were defined in ~/.zshrc for plugin ($plugins); do if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then source $ZSH/plugins/$plugin/$plugin.plugin.zsh fi done # Load the theme source "$ZSH/themes/$ZSH_THEME.zsh-theme" # Check for updates on initial load... if [ "$DISABLE_AUTO_UPDATE" = "true" ] then return else /usr/bin/env zsh $ZSH/tools/check_for_upgrade.sh fi <|file_sep|>updated/oh-my-zsh.sh
autoload -U compinit compinit -i # Load all of the plugins that were defined in ~/.zshrc for plugin ($plugins); do if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then source $ZSH/plugins/$plugin/$plugin.plugin.zsh fi done # Load all of your custom configurations from custom/ for config_file ($ZSH/custom/*.zsh) source $config_file # Load the theme source "$ZSH/themes/$ZSH_THEME.zsh-theme" # Check for updates on initial load... if [ "$DISABLE_AUTO_UPDATE" = "true" ] then return else
<|file_sep|>oh-my-zsh.sh.diff original: # Load all of your custom configurations from custom/ for config_file ($ZSH/custom/*.zsh) source $config_file updated: <|file_sep|>original/oh-my-zsh.sh # Load and run compinit autoload -U compinit compinit -i # Load all of your custom configurations from custom/ for config_file ($ZSH/custom/*.zsh) source $config_file # Load all of the plugins that were defined in ~/.zshrc for plugin ($plugins); do if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then source $ZSH/plugins/$plugin/$plugin.plugin.zsh fi done # Load the theme source "$ZSH/themes/$ZSH_THEME.zsh-theme" # Check for updates on initial load... if [ "$DISABLE_AUTO_UPDATE" = "true" ] then return <|file_sep|>current/oh-my-zsh.sh # Load and run compinit autoload -U compinit compinit -i # Load all of the plugins that were defined in ~/.zshrc for plugin ($plugins); do if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then source $ZSH/plugins/$plugin/$plugin.plugin.zsh fi done # Load the theme source "$ZSH/themes/$ZSH_THEME.zsh-theme" # Check for updates on initial load... if [ "$DISABLE_AUTO_UPDATE" = "true" ] then return else /usr/bin/env zsh $ZSH/tools/check_for_upgrade.sh fi <|file_sep|>updated/oh-my-zsh.sh autoload -U compinit compinit -i # Load all of the plugins that were defined in ~/.zshrc for plugin ($plugins); do if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then source $ZSH/plugins/$plugin/$plugin.plugin.zsh fi done # Load all of your custom configurations from custom/ for config_file ($ZSH/custom/*.zsh) source $config_file # Load the theme source "$ZSH/themes/$ZSH_THEME.zsh-theme" # Check for updates on initial load... if [ "$DISABLE_AUTO_UPDATE" = "true" ] then return else
58a7cf19b7af136c5f59ee706ffe4772871f00fa
oh-my-zsh.sh
oh-my-zsh.sh
Shell
<|file_sep|>kitchen.yml.diff original: - name: ubuntu-14.04 updated: - name: centos-8 <|file_sep|>original/kitchen.yml environments_path: test/fixtures/environments data_bags_path: test/fixtures/data_bags client_rb: environment: test platforms: - name: centos-6 - name: centos-7 - name: ubuntu-14.04 - name: ubuntu-16.04 suites: - name: default run_list: - recipe[chef-vault] - recipe[test] - name: secret_resource run_list: - recipe[chef-vault] - recipe[test::chef_vault_secret] - name: with-sugar <|file_sep|>current/kitchen.yml environments_path: test/fixtures/environments data_bags_path: test/fixtures/data_bags client_rb: environment: test platforms: - name: centos-6 - name: centos-7 - name: centos-8 - name: ubuntu-16.04 suites: - name: default run_list: - recipe[chef-vault] - recipe[test] - name: secret_resource run_list: - recipe[chef-vault] - recipe[test::chef_vault_secret] - name: with-sugar <|file_sep|>updated/kitchen.yml
environments_path: test/fixtures/environments data_bags_path: test/fixtures/data_bags client_rb: environment: test platforms: - name: centos-6 - name: centos-7 - name: centos-8 - name: ubuntu-16.04 - name: ubuntu-18.04 suites: - name: default run_list: - recipe[chef-vault] - recipe[test] - name: secret_resource run_list: - recipe[chef-vault] - recipe[test::chef_vault_secret]
<|file_sep|>kitchen.yml.diff original: - name: ubuntu-14.04 updated: - name: centos-8 <|file_sep|>original/kitchen.yml environments_path: test/fixtures/environments data_bags_path: test/fixtures/data_bags client_rb: environment: test platforms: - name: centos-6 - name: centos-7 - name: ubuntu-14.04 - name: ubuntu-16.04 suites: - name: default run_list: - recipe[chef-vault] - recipe[test] - name: secret_resource run_list: - recipe[chef-vault] - recipe[test::chef_vault_secret] - name: with-sugar <|file_sep|>current/kitchen.yml environments_path: test/fixtures/environments data_bags_path: test/fixtures/data_bags client_rb: environment: test platforms: - name: centos-6 - name: centos-7 - name: centos-8 - name: ubuntu-16.04 suites: - name: default run_list: - recipe[chef-vault] - recipe[test] - name: secret_resource run_list: - recipe[chef-vault] - recipe[test::chef_vault_secret] - name: with-sugar <|file_sep|>updated/kitchen.yml environments_path: test/fixtures/environments data_bags_path: test/fixtures/data_bags client_rb: environment: test platforms: - name: centos-6 - name: centos-7 - name: centos-8 - name: ubuntu-16.04 - name: ubuntu-18.04 suites: - name: default run_list: - recipe[chef-vault] - recipe[test] - name: secret_resource run_list: - recipe[chef-vault] - recipe[test::chef_vault_secret]
cbd6ab722422548f0ce8a4436831f9c6a23310cb
kitchen.yml
kitchen.yml
YAML
<|file_sep|>original/setup/shells.sh #!/bin/sh . ./utils.sh || exit 1 # add shells to /etc/shells shells="zsh bash dash fish ksh pdksh xonsh" for shell in $shells; do _test_executable "$shell" || return 1 shell_path="$(command -v "$shell")" if grep -Fqx "$shell_path" /etc/shells; then printf "Shell %s is already installed\n" "$shell_path" else printf "Adding shell %s to /etc/shells\n" "$shell_path" printf "%s\n" "$shell_path" \ | sudo tee -a /etc/shells >/dev/null fi done # switch default shell to zsh # _test_executable "zsh" && chsh -s "$(command -v zsh)" <|file_sep|>current/setup/shells.sh #!/bin/sh . ./utils.sh || exit 1 # add shells to /etc/shells shells="zsh bash dash fish ksh pdksh xonsh" for shell in $shells; do _test_executable "$shell" || return 1 shell_path="$(command -v "$shell")" if grep -Fqx "$shell_path" /etc/shells; then printf "Shell %s is already installed\n" "$shell_path" else printf "Adding shell %s to /etc/shells\n" "$shell_path" printf "%s\n" "$shell_path" \ | sudo tee -a /etc/shells >/dev/null fi done # switch default shell to zsh # _test_executable "zsh" && chsh -s "$(command -v zsh)" <|file_sep|>updated/setup/shells.sh
#!/bin/sh . ./utils.sh || exit 1 # add shells to /etc/shells shells="zsh bash dash fish ksh pdksh xonsh" for shell in $shells; do _test_executable "$shell" || return 1 shell_path="$(command -v "$shell")" if grep -Fqx "$shell_path" /etc/shells; then printf "Shell %s is already installed\n" "$shell_path" else printf "Adding shell %s to /etc/shells\n" "$shell_path" printf "%s\n" "$shell_path" \ | sudo tee -a /etc/shells >/dev/null fi done # switch default shell to zsh _test_executable "zsh" 2>/dev/null && chsh -s "$(command -v zsh)"
<|file_sep|>original/setup/shells.sh #!/bin/sh . ./utils.sh || exit 1 # add shells to /etc/shells shells="zsh bash dash fish ksh pdksh xonsh" for shell in $shells; do _test_executable "$shell" || return 1 shell_path="$(command -v "$shell")" if grep -Fqx "$shell_path" /etc/shells; then printf "Shell %s is already installed\n" "$shell_path" else printf "Adding shell %s to /etc/shells\n" "$shell_path" printf "%s\n" "$shell_path" \ | sudo tee -a /etc/shells >/dev/null fi done # switch default shell to zsh # _test_executable "zsh" && chsh -s "$(command -v zsh)" <|file_sep|>current/setup/shells.sh #!/bin/sh . ./utils.sh || exit 1 # add shells to /etc/shells shells="zsh bash dash fish ksh pdksh xonsh" for shell in $shells; do _test_executable "$shell" || return 1 shell_path="$(command -v "$shell")" if grep -Fqx "$shell_path" /etc/shells; then printf "Shell %s is already installed\n" "$shell_path" else printf "Adding shell %s to /etc/shells\n" "$shell_path" printf "%s\n" "$shell_path" \ | sudo tee -a /etc/shells >/dev/null fi done # switch default shell to zsh # _test_executable "zsh" && chsh -s "$(command -v zsh)" <|file_sep|>updated/setup/shells.sh #!/bin/sh . ./utils.sh || exit 1 # add shells to /etc/shells shells="zsh bash dash fish ksh pdksh xonsh" for shell in $shells; do _test_executable "$shell" || return 1 shell_path="$(command -v "$shell")" if grep -Fqx "$shell_path" /etc/shells; then printf "Shell %s is already installed\n" "$shell_path" else printf "Adding shell %s to /etc/shells\n" "$shell_path" printf "%s\n" "$shell_path" \ | sudo tee -a /etc/shells >/dev/null fi done # switch default shell to zsh _test_executable "zsh" 2>/dev/null && chsh -s "$(command -v zsh)"
796ca0bf40baa76320a07ca6525fc60d0aff53c0
setup/shells.sh
setup/shells.sh
Shell
<|file_sep|>original/requirements/base.txt dj-database-url==0.5.0 django-apiblueprint-view==1.1.2 django-basicauth==0.5.1 django-extensions==2.1.3 django-localflavor==2.1 django-markdown-deux==1.0.5 django==1.11.16 # pyup: >=1.11,<2.0 djangorestframework==3.7.7 djangorestframework-gis==0.12.0 django-cors-headers==2.4.0 fastkml==0.11 lxml==4.2.5 psycopg2-binary==2.7.5 pyshp==1.2.12 requests==2.20.0 boto==2.49.0 uk-geo-utils==0.6.0 git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7 git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1 <|file_sep|>current/requirements/base.txt dj-database-url==0.5.0 django-apiblueprint-view==1.1.2 django-basicauth==0.5.1 django-extensions==2.1.3 django-localflavor==2.1 django-markdown-deux==1.0.5 django==1.11.16 # pyup: >=1.11,<2.0 djangorestframework==3.7.7 djangorestframework-gis==0.12.0 django-cors-headers==2.4.0 fastkml==0.11 lxml==4.2.5 psycopg2-binary==2.7.5 pyshp==1.2.12 requests==2.20.0 boto==2.49.0 uk-geo-utils==0.6.0 git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7 git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1 <|file_sep|>updated/requirements/base.txt
dj-database-url==0.5.0 django-apiblueprint-view==2.0.0 django-basicauth==0.5.1 django-extensions==2.1.3 django-localflavor==2.1 django-markdown-deux==1.0.5 django==1.11.16 # pyup: >=1.11,<2.0 djangorestframework==3.7.7 djangorestframework-gis==0.12.0 django-cors-headers==2.4.0 fastkml==0.11 lxml==4.2.5 psycopg2-binary==2.7.5 pyshp==1.2.12 requests==2.20.0 boto==2.49.0 uk-geo-utils==0.6.0 git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7 git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1
<|file_sep|>original/requirements/base.txt dj-database-url==0.5.0 django-apiblueprint-view==1.1.2 django-basicauth==0.5.1 django-extensions==2.1.3 django-localflavor==2.1 django-markdown-deux==1.0.5 django==1.11.16 # pyup: >=1.11,<2.0 djangorestframework==3.7.7 djangorestframework-gis==0.12.0 django-cors-headers==2.4.0 fastkml==0.11 lxml==4.2.5 psycopg2-binary==2.7.5 pyshp==1.2.12 requests==2.20.0 boto==2.49.0 uk-geo-utils==0.6.0 git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7 git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1 <|file_sep|>current/requirements/base.txt dj-database-url==0.5.0 django-apiblueprint-view==1.1.2 django-basicauth==0.5.1 django-extensions==2.1.3 django-localflavor==2.1 django-markdown-deux==1.0.5 django==1.11.16 # pyup: >=1.11,<2.0 djangorestframework==3.7.7 djangorestframework-gis==0.12.0 django-cors-headers==2.4.0 fastkml==0.11 lxml==4.2.5 psycopg2-binary==2.7.5 pyshp==1.2.12 requests==2.20.0 boto==2.49.0 uk-geo-utils==0.6.0 git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7 git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1 <|file_sep|>updated/requirements/base.txt dj-database-url==0.5.0 django-apiblueprint-view==2.0.0 django-basicauth==0.5.1 django-extensions==2.1.3 django-localflavor==2.1 django-markdown-deux==1.0.5 django==1.11.16 # pyup: >=1.11,<2.0 djangorestframework==3.7.7 djangorestframework-gis==0.12.0 django-cors-headers==2.4.0 fastkml==0.11 lxml==4.2.5 psycopg2-binary==2.7.5 pyshp==1.2.12 requests==2.20.0 boto==2.49.0 uk-geo-utils==0.6.0 git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7 git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1
afdc86a98581935edd45b94a9d754d2dddac030d
requirements/base.txt
requirements/base.txt
Text
<|file_sep|>original/requirements-dev.txt coverage cython>=0.18 mock>=1.0 nose>=1.2 numpy https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz <|file_sep|>current/requirements-dev.txt coverage cython>=0.18 mock>=1.0 nose>=1.2 numpy https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz <|file_sep|>updated/requirements-dev.txt
coverage cython>=0.18 mock>=1.0 nose==1.2.1 numpy https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz
<|file_sep|>original/requirements-dev.txt coverage cython>=0.18 mock>=1.0 nose>=1.2 numpy https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz <|file_sep|>current/requirements-dev.txt coverage cython>=0.18 mock>=1.0 nose>=1.2 numpy https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz <|file_sep|>updated/requirements-dev.txt coverage cython>=0.18 mock>=1.0 nose==1.2.1 numpy https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz
05a4cc53dad9360bfa643e34f113697c52e3ad54
requirements-dev.txt
requirements-dev.txt
Text
<|file_sep|>src/Allegro/C.hsc.diff original: foreign import ccall "al_install_system" updated: foreign import ccall unsafe "al_install_system" <|file_sep|>original/src/Allegro/C.hsc {-# LANGUAGE ForeignFunctionInterface #-} module Allegro.C (al_init) where import Foreign import Foreign.C.Types #include <allegro5/allegro.h> al_init :: IO Bool al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit foreign import ccall "al_install_system" al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool foreign import ccall "&atexit" fp_atexit :: FunPtr (Ptr () -> IO CInt) <|file_sep|>current/src/Allegro/C.hsc {-# LANGUAGE ForeignFunctionInterface #-} module Allegro.C (al_init) where import Foreign import Foreign.C.Types #include <allegro5/allegro.h> al_init :: IO Bool al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit foreign import ccall unsafe "al_install_system" al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool foreign import ccall "&atexit" fp_atexit :: FunPtr (Ptr () -> IO CInt) <|file_sep|>updated/src/Allegro/C.hsc
{-# LANGUAGE ForeignFunctionInterface #-} module Allegro.C (al_init) where import Foreign import Foreign.C.Types #include <allegro5/allegro.h> al_init :: IO Bool al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit foreign import ccall unsafe "al_install_system" al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool foreign import ccall unsafe "&atexit" fp_atexit :: FunPtr (Ptr () -> IO CInt)
<|file_sep|>src/Allegro/C.hsc.diff original: foreign import ccall "al_install_system" updated: foreign import ccall unsafe "al_install_system" <|file_sep|>original/src/Allegro/C.hsc {-# LANGUAGE ForeignFunctionInterface #-} module Allegro.C (al_init) where import Foreign import Foreign.C.Types #include <allegro5/allegro.h> al_init :: IO Bool al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit foreign import ccall "al_install_system" al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool foreign import ccall "&atexit" fp_atexit :: FunPtr (Ptr () -> IO CInt) <|file_sep|>current/src/Allegro/C.hsc {-# LANGUAGE ForeignFunctionInterface #-} module Allegro.C (al_init) where import Foreign import Foreign.C.Types #include <allegro5/allegro.h> al_init :: IO Bool al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit foreign import ccall unsafe "al_install_system" al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool foreign import ccall "&atexit" fp_atexit :: FunPtr (Ptr () -> IO CInt) <|file_sep|>updated/src/Allegro/C.hsc {-# LANGUAGE ForeignFunctionInterface #-} module Allegro.C (al_init) where import Foreign import Foreign.C.Types #include <allegro5/allegro.h> al_init :: IO Bool al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit foreign import ccall unsafe "al_install_system" al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool foreign import ccall unsafe "&atexit" fp_atexit :: FunPtr (Ptr () -> IO CInt)
87adce0d38a35398ee200484d02d2975a6b4126e
src/Allegro/C.hsc
src/Allegro/C.hsc
Haskell
<|file_sep|>metadata/mf.asciitext.lite.yml.diff original: updated: - versionName: 2.4.0 versionCode: 13 commit: 2.4.0 subdir: base/ gradle: - yes <|file_sep|>original/metadata/mf.asciitext.lite.yml AuthorWebSite: https://oss.mobilefirst.me/ SourceCode: https://github.com/MobileFirstLLC/irregular-expressions IssueTracker: https://github.com/MobileFirstLLC/irregular-expressions/issues AutoName: Irregular Expressions RepoType: git Repo: https://github.com/MobileFirstLLC/irregular-expressions Builds: - versionName: 2.3.0 versionCode: 12 commit: 2.3.0 subdir: base/ gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 2.3.0 CurrentVersionCode: 12 <|file_sep|>current/metadata/mf.asciitext.lite.yml Repo: https://github.com/MobileFirstLLC/irregular-expressions Builds: - versionName: 2.3.0 versionCode: 12 commit: 2.3.0 subdir: base/ gradle: - yes - versionName: 2.4.0 versionCode: 13 commit: 2.4.0 subdir: base/ gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 2.3.0 CurrentVersionCode: 12 <|file_sep|>updated/metadata/mf.asciitext.lite.yml
Repo: https://github.com/MobileFirstLLC/irregular-expressions Builds: - versionName: 2.3.0 versionCode: 12 commit: 2.3.0 subdir: base/ gradle: - yes - versionName: 2.4.0 versionCode: 13 commit: 2.4.0 subdir: base/ gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 2.4.0 CurrentVersionCode: 13
<|file_sep|>metadata/mf.asciitext.lite.yml.diff original: updated: - versionName: 2.4.0 versionCode: 13 commit: 2.4.0 subdir: base/ gradle: - yes <|file_sep|>original/metadata/mf.asciitext.lite.yml AuthorWebSite: https://oss.mobilefirst.me/ SourceCode: https://github.com/MobileFirstLLC/irregular-expressions IssueTracker: https://github.com/MobileFirstLLC/irregular-expressions/issues AutoName: Irregular Expressions RepoType: git Repo: https://github.com/MobileFirstLLC/irregular-expressions Builds: - versionName: 2.3.0 versionCode: 12 commit: 2.3.0 subdir: base/ gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 2.3.0 CurrentVersionCode: 12 <|file_sep|>current/metadata/mf.asciitext.lite.yml Repo: https://github.com/MobileFirstLLC/irregular-expressions Builds: - versionName: 2.3.0 versionCode: 12 commit: 2.3.0 subdir: base/ gradle: - yes - versionName: 2.4.0 versionCode: 13 commit: 2.4.0 subdir: base/ gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 2.3.0 CurrentVersionCode: 12 <|file_sep|>updated/metadata/mf.asciitext.lite.yml Repo: https://github.com/MobileFirstLLC/irregular-expressions Builds: - versionName: 2.3.0 versionCode: 12 commit: 2.3.0 subdir: base/ gradle: - yes - versionName: 2.4.0 versionCode: 13 commit: 2.4.0 subdir: base/ gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 2.4.0 CurrentVersionCode: 13
e60cec9826471920ab393bfe386f2b3dc36206a8
metadata/mf.asciitext.lite.yml
metadata/mf.asciitext.lite.yml
YAML
<|file_sep|>spec/api-system-preferences-spec.js.diff original: if (process.platform !== 'darwin') { return } updated: describe('systemPreferences.getAccentColor', function () { if (process.platform !== 'win32') { return } it('should return a non-empty string', function () { let accentColor = systemPreferences.getAccentColor(); assert.notEqual(accentColor, null); assert(accentColor.length > 0); }) }) <|file_sep|>original/spec/api-system-preferences-spec.js const assert = require('assert') const {remote} = require('electron') const {systemPreferences} = remote describe('systemPreferences module', function () { if (process.platform !== 'darwin') { return } describe('systemPreferences.getUserDefault(key, type)', function () { it('returns values for known user defaults', function () { let locale = systemPreferences.getUserDefault('AppleLocale', 'string') assert.notEqual(locale, null) assert(locale.length > 0) let languages = systemPreferences.getUserDefault('AppleLanguages', 'array') assert.notEqual(languages, null) assert(languages.length > 0) }) }) }) <|file_sep|>current/spec/api-system-preferences-spec.js } it('should return a non-empty string', function () { let accentColor = systemPreferences.getAccentColor(); assert.notEqual(accentColor, null); assert(accentColor.length > 0); }) }) describe('systemPreferences.getUserDefault(key, type)', function () { it('returns values for known user defaults', function () { let locale = systemPreferences.getUserDefault('AppleLocale', 'string') assert.notEqual(locale, null) assert(locale.length > 0) let languages = systemPreferences.getUserDefault('AppleLanguages', 'array') assert.notEqual(languages, null) assert(languages.length > 0) }) }) }) <|file_sep|>updated/spec/api-system-preferences-spec.js
} it('should return a non-empty string', function () { let accentColor = systemPreferences.getAccentColor(); assert.notEqual(accentColor, null); assert(accentColor.length > 0); }) }) describe('systemPreferences.getUserDefault(key, type)', function () { if (process.platform !== 'darwin') { return } it('returns values for known user defaults', function () { let locale = systemPreferences.getUserDefault('AppleLocale', 'string') assert.notEqual(locale, null) assert(locale.length > 0) let languages = systemPreferences.getUserDefault('AppleLanguages', 'array') assert.notEqual(languages, null)
<|file_sep|>spec/api-system-preferences-spec.js.diff original: if (process.platform !== 'darwin') { return } updated: describe('systemPreferences.getAccentColor', function () { if (process.platform !== 'win32') { return } it('should return a non-empty string', function () { let accentColor = systemPreferences.getAccentColor(); assert.notEqual(accentColor, null); assert(accentColor.length > 0); }) }) <|file_sep|>original/spec/api-system-preferences-spec.js const assert = require('assert') const {remote} = require('electron') const {systemPreferences} = remote describe('systemPreferences module', function () { if (process.platform !== 'darwin') { return } describe('systemPreferences.getUserDefault(key, type)', function () { it('returns values for known user defaults', function () { let locale = systemPreferences.getUserDefault('AppleLocale', 'string') assert.notEqual(locale, null) assert(locale.length > 0) let languages = systemPreferences.getUserDefault('AppleLanguages', 'array') assert.notEqual(languages, null) assert(languages.length > 0) }) }) }) <|file_sep|>current/spec/api-system-preferences-spec.js } it('should return a non-empty string', function () { let accentColor = systemPreferences.getAccentColor(); assert.notEqual(accentColor, null); assert(accentColor.length > 0); }) }) describe('systemPreferences.getUserDefault(key, type)', function () { it('returns values for known user defaults', function () { let locale = systemPreferences.getUserDefault('AppleLocale', 'string') assert.notEqual(locale, null) assert(locale.length > 0) let languages = systemPreferences.getUserDefault('AppleLanguages', 'array') assert.notEqual(languages, null) assert(languages.length > 0) }) }) }) <|file_sep|>updated/spec/api-system-preferences-spec.js } it('should return a non-empty string', function () { let accentColor = systemPreferences.getAccentColor(); assert.notEqual(accentColor, null); assert(accentColor.length > 0); }) }) describe('systemPreferences.getUserDefault(key, type)', function () { if (process.platform !== 'darwin') { return } it('returns values for known user defaults', function () { let locale = systemPreferences.getUserDefault('AppleLocale', 'string') assert.notEqual(locale, null) assert(locale.length > 0) let languages = systemPreferences.getUserDefault('AppleLanguages', 'array') assert.notEqual(languages, null)
9f946527929d65607c0d42d33733d370fee27f83
spec/api-system-preferences-spec.js
spec/api-system-preferences-spec.js
JavaScript
<|file_sep|>original/config/routes.rb end scope :path => Refinery::Products.products_categories_path do resources :categories, :path => '', :as => :categories, :controller => 'categories' end end # Admin routes namespace :products, :path => '' do namespace :admin, :path => Refinery::Core.backend_route do scope :path => Refinery::Products.products_path do root :to => "products#index" resources :products, :except => :show do collection do post :update_positions get :uncategorized end end resources :categories <|file_sep|>current/config/routes.rb end scope :path => Refinery::Products.products_categories_path do resources :categories, :path => '', :as => :categories, :controller => 'categories' end end # Admin routes namespace :products, :path => '' do namespace :admin, :path => Refinery::Core.backend_route do scope :path => Refinery::Products.products_path do root :to => "products#index" resources :products, :except => :show do collection do post :update_positions get :uncategorized end end resources :categories <|file_sep|>updated/config/routes.rb
end scope :path => Refinery::Products.products_categories_path do resources :categories, :path => '', :as => :categories, :controller => 'categories' end end # Admin routes namespace :products, :path => '' do namespace :admin, :path => Refinery::Core.backend_route do scope :path => Refinery::Products.shop_path do root :to => "products#index" resources :products, :except => :show do collection do post :update_positions get :uncategorized end end resources :categories
<|file_sep|>original/config/routes.rb end scope :path => Refinery::Products.products_categories_path do resources :categories, :path => '', :as => :categories, :controller => 'categories' end end # Admin routes namespace :products, :path => '' do namespace :admin, :path => Refinery::Core.backend_route do scope :path => Refinery::Products.products_path do root :to => "products#index" resources :products, :except => :show do collection do post :update_positions get :uncategorized end end resources :categories <|file_sep|>current/config/routes.rb end scope :path => Refinery::Products.products_categories_path do resources :categories, :path => '', :as => :categories, :controller => 'categories' end end # Admin routes namespace :products, :path => '' do namespace :admin, :path => Refinery::Core.backend_route do scope :path => Refinery::Products.products_path do root :to => "products#index" resources :products, :except => :show do collection do post :update_positions get :uncategorized end end resources :categories <|file_sep|>updated/config/routes.rb end scope :path => Refinery::Products.products_categories_path do resources :categories, :path => '', :as => :categories, :controller => 'categories' end end # Admin routes namespace :products, :path => '' do namespace :admin, :path => Refinery::Core.backend_route do scope :path => Refinery::Products.shop_path do root :to => "products#index" resources :products, :except => :show do collection do post :update_positions get :uncategorized end end resources :categories
bfdfd0b50ad78a31a93485b19e74b1a83e11e51a
config/routes.rb
config/routes.rb
Ruby
<|file_sep|>original/{{cookiecutter.project_slug}}/requirements/base.txt Django==1.11.1 wagtail==1.10.1 wagtailfontawesome==1.0.6 celery==4.0.2 django-cachalot==1.4.1 django-compressor==2.1.1 django-environ==0.4.3 django-foundation-formtags==0.0.8 django-wagtail-feeds==0.0.7 django-libsass==0.7 django-redis==4.8.0 libsass==0.12.3 psycopg2==2.7.1 wand==0.4.4 {% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %} {% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %} <|file_sep|>current/{{cookiecutter.project_slug}}/requirements/base.txt Django==1.11.1 wagtail==1.10.1 wagtailfontawesome==1.0.6 celery==4.0.2 django-cachalot==1.4.1 django-compressor==2.1.1 django-environ==0.4.3 django-foundation-formtags==0.0.8 django-wagtail-feeds==0.0.7 django-libsass==0.7 django-redis==4.8.0 libsass==0.12.3 psycopg2==2.7.1 wand==0.4.4 {% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %} {% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %} <|file_sep|>updated/{{cookiecutter.project_slug}}/requirements/base.txt
Django==1.11.1 wagtail==1.10.1 wagtailfontawesome==1.0.6 celery==4.0.2 django-cachalot==1.5.0 django-compressor==2.1.1 django-environ==0.4.3 django-foundation-formtags==0.0.8 django-wagtail-feeds==0.0.7 django-libsass==0.7 django-redis==4.8.0 libsass==0.12.3 psycopg2==2.7.1 wand==0.4.4 {% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %} {% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %}
<|file_sep|>original/{{cookiecutter.project_slug}}/requirements/base.txt Django==1.11.1 wagtail==1.10.1 wagtailfontawesome==1.0.6 celery==4.0.2 django-cachalot==1.4.1 django-compressor==2.1.1 django-environ==0.4.3 django-foundation-formtags==0.0.8 django-wagtail-feeds==0.0.7 django-libsass==0.7 django-redis==4.8.0 libsass==0.12.3 psycopg2==2.7.1 wand==0.4.4 {% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %} {% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %} <|file_sep|>current/{{cookiecutter.project_slug}}/requirements/base.txt Django==1.11.1 wagtail==1.10.1 wagtailfontawesome==1.0.6 celery==4.0.2 django-cachalot==1.4.1 django-compressor==2.1.1 django-environ==0.4.3 django-foundation-formtags==0.0.8 django-wagtail-feeds==0.0.7 django-libsass==0.7 django-redis==4.8.0 libsass==0.12.3 psycopg2==2.7.1 wand==0.4.4 {% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %} {% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %} <|file_sep|>updated/{{cookiecutter.project_slug}}/requirements/base.txt Django==1.11.1 wagtail==1.10.1 wagtailfontawesome==1.0.6 celery==4.0.2 django-cachalot==1.5.0 django-compressor==2.1.1 django-environ==0.4.3 django-foundation-formtags==0.0.8 django-wagtail-feeds==0.0.7 django-libsass==0.7 django-redis==4.8.0 libsass==0.12.3 psycopg2==2.7.1 wand==0.4.4 {% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %} {% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %}
cdeb40250439322fb444633545abef3c0e82eb61
{{cookiecutter.project_slug}}/requirements/base.txt
{{cookiecutter.project_slug}}/requirements/base.txt
Text
<|file_sep|>original/package.json { "name": "cloudcp", "version": "0.1.0", "description": "Copy local files to any cloud storage provider supported by pkgcloud", "main": "index.js", "scripts": { "test": "tap test/index.js", "test-ci": "standard && istanbul cover test/index.js", "preversion": "npm test", "postversion": "git push && git push --tags" }, "keywords": [ "copy", "fs", "cp", "cloud" ], "author": "Pier-Luc Gendreau", "license": "MIT", "devDependencies": { "glob": "^5.0.15", <|file_sep|>current/package.json { "name": "cloudcp", "version": "0.1.0", "description": "Copy local files to any cloud storage provider supported by pkgcloud", "main": "index.js", "scripts": { "test": "tap test/index.js", "test-ci": "standard && istanbul cover test/index.js", "preversion": "npm test", "postversion": "git push && git push --tags" }, "keywords": [ "copy", "fs", "cp", "cloud" ], "author": "Pier-Luc Gendreau", "license": "MIT", "devDependencies": { "glob": "^5.0.15", <|file_sep|>updated/package.json
{ "name": "cloudcp", "version": "0.1.0", "description": "Copy local files to any cloud storage provider supported by pkgcloud", "main": "index.js", "scripts": { "test": "tap test/index.js", "test-ci": "standard && istanbul cover test/index.js" }, "keywords": [ "copy", "fs", "cp", "cloud" ], "author": "Pier-Luc Gendreau", "license": "MIT", "devDependencies": { "glob": "^5.0.15", "istanbul": "^0.4.0", "sinon": "^1.17.2",
<|file_sep|>original/package.json { "name": "cloudcp", "version": "0.1.0", "description": "Copy local files to any cloud storage provider supported by pkgcloud", "main": "index.js", "scripts": { "test": "tap test/index.js", "test-ci": "standard && istanbul cover test/index.js", "preversion": "npm test", "postversion": "git push && git push --tags" }, "keywords": [ "copy", "fs", "cp", "cloud" ], "author": "Pier-Luc Gendreau", "license": "MIT", "devDependencies": { "glob": "^5.0.15", <|file_sep|>current/package.json { "name": "cloudcp", "version": "0.1.0", "description": "Copy local files to any cloud storage provider supported by pkgcloud", "main": "index.js", "scripts": { "test": "tap test/index.js", "test-ci": "standard && istanbul cover test/index.js", "preversion": "npm test", "postversion": "git push && git push --tags" }, "keywords": [ "copy", "fs", "cp", "cloud" ], "author": "Pier-Luc Gendreau", "license": "MIT", "devDependencies": { "glob": "^5.0.15", <|file_sep|>updated/package.json { "name": "cloudcp", "version": "0.1.0", "description": "Copy local files to any cloud storage provider supported by pkgcloud", "main": "index.js", "scripts": { "test": "tap test/index.js", "test-ci": "standard && istanbul cover test/index.js" }, "keywords": [ "copy", "fs", "cp", "cloud" ], "author": "Pier-Luc Gendreau", "license": "MIT", "devDependencies": { "glob": "^5.0.15", "istanbul": "^0.4.0", "sinon": "^1.17.2",
9fce8b4849a712e559929feac34cc5c1fa8b6c76
package.json
package.json
JSON
<|file_sep|>plugins/cache_reload.js.diff original: url: "", initialized: false updated: url: "" <|file_sep|>plugins/cache_reload.js.diff original: if(this.initalized) updated: if(!impl.url) <|file_sep|>original/plugins/cache_reload.js if(this.initalized) return this; BOOMR.subscribe( "page_ready", function() { if(!impl.url) return; // we use document and not BOOMR.window.document since // we can run inside the boomerang iframe if any var i=document.createElement('iframe'); i.style.display="none"; i.src=impl.url; document.body.appendChild(i); }, null, null ); this.initialized = true; <|file_sep|>current/plugins/cache_reload.js if(!impl.url) return this; BOOMR.subscribe( "page_ready", function() { if(!impl.url) return; // we use document and not BOOMR.window.document since // we can run inside the boomerang iframe if any var i=document.createElement('iframe'); i.style.display="none"; i.src=impl.url; document.body.appendChild(i); }, null, null ); this.initialized = true; return this; <|file_sep|>updated/plugins/cache_reload.js
if(!impl.url) return this; // we use document and not BOOMR.window.document since // we can run inside the boomerang iframe if any var i=document.createElement('iframe'); i.style.display="none"; i.src=impl.url; document.body.appendChild(i); return this; }, is_complete: function() { // we always return true since this plugin never adds anything to the beacon return true; } }; }());
<|file_sep|>plugins/cache_reload.js.diff original: url: "", initialized: false updated: url: "" <|file_sep|>plugins/cache_reload.js.diff original: if(this.initalized) updated: if(!impl.url) <|file_sep|>original/plugins/cache_reload.js if(this.initalized) return this; BOOMR.subscribe( "page_ready", function() { if(!impl.url) return; // we use document and not BOOMR.window.document since // we can run inside the boomerang iframe if any var i=document.createElement('iframe'); i.style.display="none"; i.src=impl.url; document.body.appendChild(i); }, null, null ); this.initialized = true; <|file_sep|>current/plugins/cache_reload.js if(!impl.url) return this; BOOMR.subscribe( "page_ready", function() { if(!impl.url) return; // we use document and not BOOMR.window.document since // we can run inside the boomerang iframe if any var i=document.createElement('iframe'); i.style.display="none"; i.src=impl.url; document.body.appendChild(i); }, null, null ); this.initialized = true; return this; <|file_sep|>updated/plugins/cache_reload.js if(!impl.url) return this; // we use document and not BOOMR.window.document since // we can run inside the boomerang iframe if any var i=document.createElement('iframe'); i.style.display="none"; i.src=impl.url; document.body.appendChild(i); return this; }, is_complete: function() { // we always return true since this plugin never adds anything to the beacon return true; } }; }());
60a2b50f5c49c3c109f3f90e53897a1601bdd256
plugins/cache_reload.js
plugins/cache_reload.js
JavaScript
<|file_sep|>original/.travis.yml before_install: - node -v - npm -v install: - npm install - npm install -g grunt-cli@0.1.13 - grunt build - mvn install -DskipTests=true script: - mvn test after_success: - git config --global user.email "simon@simondean.org" - git config --global user.name "Simon Dean" - mvn --batch-mode release:prepare release:perform --settings settings.xml - git log -5 env: secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s= <|file_sep|>current/.travis.yml before_install: - node -v - npm -v install: - npm install - npm install -g grunt-cli@0.1.13 - grunt build - mvn install -DskipTests=true script: - mvn test after_success: - git config --global user.email "simon@simondean.org" - git config --global user.name "Simon Dean" - mvn --batch-mode release:prepare release:perform --settings settings.xml - git log -5 env: secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s= <|file_sep|>updated/.travis.yml
language: java jdk: - oraclejdk8 before_install: - node -v - npm -v install: - npm install - npm install -g grunt-cli@0.1.13 - grunt build - mvn install -DskipTests=true script: - mvn test - mvn --batch-mode deploy -DskipTests=true --settings settings.xml env: secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s=
<|file_sep|>original/.travis.yml before_install: - node -v - npm -v install: - npm install - npm install -g grunt-cli@0.1.13 - grunt build - mvn install -DskipTests=true script: - mvn test after_success: - git config --global user.email "simon@simondean.org" - git config --global user.name "Simon Dean" - mvn --batch-mode release:prepare release:perform --settings settings.xml - git log -5 env: secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s= <|file_sep|>current/.travis.yml before_install: - node -v - npm -v install: - npm install - npm install -g grunt-cli@0.1.13 - grunt build - mvn install -DskipTests=true script: - mvn test after_success: - git config --global user.email "simon@simondean.org" - git config --global user.name "Simon Dean" - mvn --batch-mode release:prepare release:perform --settings settings.xml - git log -5 env: secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s= <|file_sep|>updated/.travis.yml language: java jdk: - oraclejdk8 before_install: - node -v - npm -v install: - npm install - npm install -g grunt-cli@0.1.13 - grunt build - mvn install -DskipTests=true script: - mvn test - mvn --batch-mode deploy -DskipTests=true --settings settings.xml env: secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s=
ce5fd980f05a32a21b1a1747c3864c4ede8e4a97
.travis.yml
.travis.yml
YAML
<|file_sep|>original/manifest.json { "manifest_version": 2, "name": "AutoFlix", "version": "1.1.3", "description": "Simplifies the access to Netflix for all users.", "icons": { "16": "icon-16.png", "32": "icon-32.png", "48": "icon-48.png", "64": "icon-64.png" , "128": "icon-128.png" }, "browser_action": { "default_icon": "icon-128.png" }, "background": { "scripts": ["background.js"] <|file_sep|>current/manifest.json { "manifest_version": 2, "name": "AutoFlix", "version": "1.1.3", "description": "Simplifies the access to Netflix for all users.", "icons": { "16": "icon-16.png", "32": "icon-32.png", "48": "icon-48.png", "64": "icon-64.png" , "128": "icon-128.png" }, "browser_action": { "default_icon": "icon-128.png" }, "background": { "scripts": ["background.js"] <|file_sep|>updated/manifest.json
{ "manifest_version": 2, "name": "AutoFlix", "version": "1.1.4", "description": "Simplifies the access to Netflix for all users.", "icons": { "16": "icon-16.png", "32": "icon-32.png", "48": "icon-48.png", "64": "icon-64.png" , "128": "icon-128.png" }, "browser_action": { "default_icon": "icon-128.png" }, "background": { "scripts": ["background.js"]
<|file_sep|>original/manifest.json { "manifest_version": 2, "name": "AutoFlix", "version": "1.1.3", "description": "Simplifies the access to Netflix for all users.", "icons": { "16": "icon-16.png", "32": "icon-32.png", "48": "icon-48.png", "64": "icon-64.png" , "128": "icon-128.png" }, "browser_action": { "default_icon": "icon-128.png" }, "background": { "scripts": ["background.js"] <|file_sep|>current/manifest.json { "manifest_version": 2, "name": "AutoFlix", "version": "1.1.3", "description": "Simplifies the access to Netflix for all users.", "icons": { "16": "icon-16.png", "32": "icon-32.png", "48": "icon-48.png", "64": "icon-64.png" , "128": "icon-128.png" }, "browser_action": { "default_icon": "icon-128.png" }, "background": { "scripts": ["background.js"] <|file_sep|>updated/manifest.json { "manifest_version": 2, "name": "AutoFlix", "version": "1.1.4", "description": "Simplifies the access to Netflix for all users.", "icons": { "16": "icon-16.png", "32": "icon-32.png", "48": "icon-48.png", "64": "icon-64.png" , "128": "icon-128.png" }, "browser_action": { "default_icon": "icon-128.png" }, "background": { "scripts": ["background.js"]
dc06a0292108738f657197352142ce459c9e14df
manifest.json
manifest.json
JSON
<|file_sep|>original/Casks/tomtom-mysports-connect.rb <|file_sep|>current/Casks/tomtom-mysports-connect.rb <|file_sep|>updated/Casks/tomtom-mysports-connect.rb
cask 'tomtom-mysports-connect' do version :latest sha256 :no_check url 'http://download.tomtom.com/sweet/fitness/MySportsConnectDesktop/current/TomTomMySportsConnectInstaller.pkg' name 'TomTom MySports Connect' homepage 'http://www.tomtom.com/de_ch/mysports/getstarted/' license :gpl pkg 'TomTomMySportsConnectInstaller.pkg' uninstall :quit => 'com.tomtom.mysportsconnect', :pkgutil => [ 'com.tomtom.tomtomfa.pkg', 'com.tomtom.tomtomfa.temp.pkg', ], :delete => '/Applications/TomTom MySports Connect.app' end
<|file_sep|>original/Casks/tomtom-mysports-connect.rb <|file_sep|>current/Casks/tomtom-mysports-connect.rb <|file_sep|>updated/Casks/tomtom-mysports-connect.rb cask 'tomtom-mysports-connect' do version :latest sha256 :no_check url 'http://download.tomtom.com/sweet/fitness/MySportsConnectDesktop/current/TomTomMySportsConnectInstaller.pkg' name 'TomTom MySports Connect' homepage 'http://www.tomtom.com/de_ch/mysports/getstarted/' license :gpl pkg 'TomTomMySportsConnectInstaller.pkg' uninstall :quit => 'com.tomtom.mysportsconnect', :pkgutil => [ 'com.tomtom.tomtomfa.pkg', 'com.tomtom.tomtomfa.temp.pkg', ], :delete => '/Applications/TomTom MySports Connect.app' end
b513cbe1839215d107b620e897e279e26dcbcb70
Casks/tomtom-mysports-connect.rb
Casks/tomtom-mysports-connect.rb
Ruby
<|file_sep|>original/.travis.yml r_packages: # - gWidgets # - gWidgetstcltk - stringdist - igraph # - covr # r packages directly installed from github r_github_pages: - jimhester/covr # before_install: # - sudo apt-get install tk after_success: - Rscript -e 'covr::codecov()' # or coveralls - Rscript -e 'covr::coveralls()' notifications: email: false <|file_sep|>current/.travis.yml r_packages: # - gWidgets # - gWidgetstcltk - stringdist - igraph # - covr # r packages directly installed from github r_github_pages: - jimhester/covr # before_install: # - sudo apt-get install tk after_success: - Rscript -e 'covr::codecov()' # or coveralls - Rscript -e 'covr::coveralls()' notifications: email: false <|file_sep|>updated/.travis.yml
# - gWidgetstcltk - stringdist - igraph # - covr # r packages directly installed from github r_github_pages: - jimhester/covr # before_install: # - sudo apt-get install tk after_success: - Rscript -e 'covr::codecov()' # or coveralls - Rscript -e 'covr::coveralls()' notifications: email: on_success: never # default: change on_failure: always # default: always
<|file_sep|>original/.travis.yml r_packages: # - gWidgets # - gWidgetstcltk - stringdist - igraph # - covr # r packages directly installed from github r_github_pages: - jimhester/covr # before_install: # - sudo apt-get install tk after_success: - Rscript -e 'covr::codecov()' # or coveralls - Rscript -e 'covr::coveralls()' notifications: email: false <|file_sep|>current/.travis.yml r_packages: # - gWidgets # - gWidgetstcltk - stringdist - igraph # - covr # r packages directly installed from github r_github_pages: - jimhester/covr # before_install: # - sudo apt-get install tk after_success: - Rscript -e 'covr::codecov()' # or coveralls - Rscript -e 'covr::coveralls()' notifications: email: false <|file_sep|>updated/.travis.yml # - gWidgetstcltk - stringdist - igraph # - covr # r packages directly installed from github r_github_pages: - jimhester/covr # before_install: # - sudo apt-get install tk after_success: - Rscript -e 'covr::codecov()' # or coveralls - Rscript -e 'covr::coveralls()' notifications: email: on_success: never # default: change on_failure: always # default: always
c8f8edb7d88164336319247459b3b11a8002e1e4
.travis.yml
.travis.yml
YAML
<|file_sep|>modules/hdo/templates/deployer.json.diff original: "command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"], updated: <|file_sep|>original/modules/hdo/templates/deployer.json { "logdir": "<%= logdir %>", "github_client_id": "<%= github_client_id %>", "github_client_secret": "<%= github_client_secret %>", "statsd": "ops1.holderdeord.no:8125", "projects": { "holderdeord/hdo-site#master": { "directory": "<%= appdir %>", "command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"], "token": "<%= token %>", "timeout": 180 } } } <|file_sep|>current/modules/hdo/templates/deployer.json { "logdir": "<%= logdir %>", "github_client_id": "<%= github_client_id %>", "github_client_secret": "<%= github_client_secret %>", "statsd": "ops1.holderdeord.no:8125", "projects": { "holderdeord/hdo-site#master": { "directory": "<%= appdir %>", "token": "<%= token %>", "timeout": 180 } } } <|file_sep|>updated/modules/hdo/templates/deployer.json
{ "logdir": "<%= logdir %>", "github_client_id": "<%= github_client_id %>", "github_client_secret": "<%= github_client_secret %>", "statsd": "ops1.holderdeord.no:8125", "projects": { "holderdeord/hdo-site#master": { "directory": "<%= appdir %>", "token": "<%= token %>", "timeout": 180, "command": [ "cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable", "deploy:cleanup" ] } }
<|file_sep|>modules/hdo/templates/deployer.json.diff original: "command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"], updated: <|file_sep|>original/modules/hdo/templates/deployer.json { "logdir": "<%= logdir %>", "github_client_id": "<%= github_client_id %>", "github_client_secret": "<%= github_client_secret %>", "statsd": "ops1.holderdeord.no:8125", "projects": { "holderdeord/hdo-site#master": { "directory": "<%= appdir %>", "command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"], "token": "<%= token %>", "timeout": 180 } } } <|file_sep|>current/modules/hdo/templates/deployer.json { "logdir": "<%= logdir %>", "github_client_id": "<%= github_client_id %>", "github_client_secret": "<%= github_client_secret %>", "statsd": "ops1.holderdeord.no:8125", "projects": { "holderdeord/hdo-site#master": { "directory": "<%= appdir %>", "token": "<%= token %>", "timeout": 180 } } } <|file_sep|>updated/modules/hdo/templates/deployer.json { "logdir": "<%= logdir %>", "github_client_id": "<%= github_client_id %>", "github_client_secret": "<%= github_client_secret %>", "statsd": "ops1.holderdeord.no:8125", "projects": { "holderdeord/hdo-site#master": { "directory": "<%= appdir %>", "token": "<%= token %>", "timeout": 180, "command": [ "cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable", "deploy:cleanup" ] } }
350ff106b2d356700eb24e8b9ceb819af263b7f4
modules/hdo/templates/deployer.json
modules/hdo/templates/deployer.json
JSON
<|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff original: updated: <|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff original: {current, commands} = case String.split(arg, ",") do [arg] -> {[arg | current], commands} # special care if the argument contains a comma args -> update_commands(args, current, commands) end gather_commands(rest, current, commands) updated: case String.split(arg, ",", parts: 2) do [arg] -> gather_commands(rest, [arg | current], commands) [left, right] -> rest = append_unless_empty(right, rest) current = append_unless_empty(left, current) gather_commands(rest, [], [current | commands]) end <|file_sep|>original/lib/mix/lib/mix/tasks/do.ex case String.split(arg, ",") do [arg] -> {[arg | current], commands} # special care if the argument contains a comma args -> update_commands(args, current, commands) end gather_commands(rest, current, commands) end defp update_commands([], current, commands) do {current, commands} end defp update_commands([arg], current, commands) when arg != "" do {[arg], [current | commands]} end defp update_commands([arg | args], current, commands) do # if the argument is empty, we had a leading or trailing comma # so we simply terminate the current command command = if arg == "", do: current, else: [arg | current] update_commands(args, [], [command | commands]) end end <|file_sep|>current/lib/mix/lib/mix/tasks/do.ex gather_commands(rest, [arg | current], commands) [left, right] -> rest = append_unless_empty(right, rest) current = append_unless_empty(left, current) gather_commands(rest, [], [current | commands]) end end defp update_commands([], current, commands) do {current, commands} end defp update_commands([arg], current, commands) when arg != "" do {[arg], [current | commands]} end defp update_commands([arg | args], current, commands) do # if the argument is empty, we had a leading or trailing comma # so we simply terminate the current command command = if arg == "", do: current, else: [arg | current] update_commands(args, [], [command | commands]) end end <|file_sep|>updated/lib/mix/lib/mix/tasks/do.ex
defp gather_commands([], current, commands) do [current | commands] |> Enum.reject(&(&1 == [])) |> Enum.map(&Enum.reverse(&1)) |> Enum.reverse end defp gather_commands([arg | rest], current, commands) do case String.split(arg, ",", parts: 2) do [arg] -> gather_commands(rest, [arg | current], commands) [left, right] -> rest = append_unless_empty(right, rest) current = append_unless_empty(left, current) gather_commands(rest, [], [current | commands]) end end defp append_unless_empty("", list), do: list defp append_unless_empty(h, list), do: [h | list] end
<|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff original: updated: <|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff original: {current, commands} = case String.split(arg, ",") do [arg] -> {[arg | current], commands} # special care if the argument contains a comma args -> update_commands(args, current, commands) end gather_commands(rest, current, commands) updated: case String.split(arg, ",", parts: 2) do [arg] -> gather_commands(rest, [arg | current], commands) [left, right] -> rest = append_unless_empty(right, rest) current = append_unless_empty(left, current) gather_commands(rest, [], [current | commands]) end <|file_sep|>original/lib/mix/lib/mix/tasks/do.ex case String.split(arg, ",") do [arg] -> {[arg | current], commands} # special care if the argument contains a comma args -> update_commands(args, current, commands) end gather_commands(rest, current, commands) end defp update_commands([], current, commands) do {current, commands} end defp update_commands([arg], current, commands) when arg != "" do {[arg], [current | commands]} end defp update_commands([arg | args], current, commands) do # if the argument is empty, we had a leading or trailing comma # so we simply terminate the current command command = if arg == "", do: current, else: [arg | current] update_commands(args, [], [command | commands]) end end <|file_sep|>current/lib/mix/lib/mix/tasks/do.ex gather_commands(rest, [arg | current], commands) [left, right] -> rest = append_unless_empty(right, rest) current = append_unless_empty(left, current) gather_commands(rest, [], [current | commands]) end end defp update_commands([], current, commands) do {current, commands} end defp update_commands([arg], current, commands) when arg != "" do {[arg], [current | commands]} end defp update_commands([arg | args], current, commands) do # if the argument is empty, we had a leading or trailing comma # so we simply terminate the current command command = if arg == "", do: current, else: [arg | current] update_commands(args, [], [command | commands]) end end <|file_sep|>updated/lib/mix/lib/mix/tasks/do.ex defp gather_commands([], current, commands) do [current | commands] |> Enum.reject(&(&1 == [])) |> Enum.map(&Enum.reverse(&1)) |> Enum.reverse end defp gather_commands([arg | rest], current, commands) do case String.split(arg, ",", parts: 2) do [arg] -> gather_commands(rest, [arg | current], commands) [left, right] -> rest = append_unless_empty(right, rest) current = append_unless_empty(left, current) gather_commands(rest, [], [current | commands]) end end defp append_unless_empty("", list), do: list defp append_unless_empty(h, list), do: [h | list] end
70ba8464c9fb2745aaeb0f53783c1361b8db982e
lib/mix/lib/mix/tasks/do.ex
lib/mix/lib/mix/tasks/do.ex
Elixir
<|file_sep|>original/examples/wfs.html height: 475px; border: 1px solid black; } </style> <script src="../lib/OpenLayers.js"></script> <script type="text/javascript"> <!-- var map, layer; function init(){ map = new OpenLayers.Map('map'); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); layer = new OpenLayers.Layer.WFS( "OpenLayers WFS", "http://octo.metacarta.com/services/search/WFS", {q: 'dolphins', maxfeatures: 10} ); map.addLayer(layer); <|file_sep|>current/examples/wfs.html height: 475px; border: 1px solid black; } </style> <script src="../lib/OpenLayers.js"></script> <script type="text/javascript"> <!-- var map, layer; function init(){ map = new OpenLayers.Map('map'); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); layer = new OpenLayers.Layer.WFS( "OpenLayers WFS", "http://octo.metacarta.com/services/search/WFS", {q: 'dolphins', maxfeatures: 10} ); map.addLayer(layer); <|file_sep|>updated/examples/wfs.html
height: 475px; border: 1px solid black; } </style> <script src="../lib/OpenLayers.js"></script> <script type="text/javascript"> <!-- var map, layer; function init(){ OpenLayers.ProxyHost = '/dev/examples/proxy.cgi?url='; map = new OpenLayers.Map('map'); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); layer = new OpenLayers.Layer.WFS( "OpenLayers WFS", "http://octo.metacarta.com/services/search/WFS", {q: 'dolphins', maxfeatures: 10} ); map.addLayer(layer);
<|file_sep|>original/examples/wfs.html height: 475px; border: 1px solid black; } </style> <script src="../lib/OpenLayers.js"></script> <script type="text/javascript"> <!-- var map, layer; function init(){ map = new OpenLayers.Map('map'); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); layer = new OpenLayers.Layer.WFS( "OpenLayers WFS", "http://octo.metacarta.com/services/search/WFS", {q: 'dolphins', maxfeatures: 10} ); map.addLayer(layer); <|file_sep|>current/examples/wfs.html height: 475px; border: 1px solid black; } </style> <script src="../lib/OpenLayers.js"></script> <script type="text/javascript"> <!-- var map, layer; function init(){ map = new OpenLayers.Map('map'); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); layer = new OpenLayers.Layer.WFS( "OpenLayers WFS", "http://octo.metacarta.com/services/search/WFS", {q: 'dolphins', maxfeatures: 10} ); map.addLayer(layer); <|file_sep|>updated/examples/wfs.html height: 475px; border: 1px solid black; } </style> <script src="../lib/OpenLayers.js"></script> <script type="text/javascript"> <!-- var map, layer; function init(){ OpenLayers.ProxyHost = '/dev/examples/proxy.cgi?url='; map = new OpenLayers.Map('map'); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); layer = new OpenLayers.Layer.WFS( "OpenLayers WFS", "http://octo.metacarta.com/services/search/WFS", {q: 'dolphins', maxfeatures: 10} ); map.addLayer(layer);
1a3d466f2a681cd5f3f9bff74a6eddf15d4a7c6b
examples/wfs.html
examples/wfs.html
HTML
<|file_sep|>original/requirements_dev.txt # Database psycopg2-binary==2.7.7 pip==19.0.1 bumpversion==0.5.3 wheel==0.32.3 watchdog==0.9.0 flake8==3.7.5 tox==3.7.0 cryptography==2.5 Sphinx==1.8.4 coverage==4.5.2 pytest==4.2.0 pytest-runner==4.2 pytest-cov==2.6.1 # miscellaneous sphinx_rtd_theme==0.4.2 pre-commit==1.14.3 <|file_sep|>current/requirements_dev.txt # Database psycopg2-binary==2.7.7 pip==19.0.1 bumpversion==0.5.3 wheel==0.32.3 watchdog==0.9.0 flake8==3.7.5 tox==3.7.0 cryptography==2.5 Sphinx==1.8.4 coverage==4.5.2 pytest==4.2.0 pytest-runner==4.2 pytest-cov==2.6.1 # miscellaneous sphinx_rtd_theme==0.4.2 pre-commit==1.14.3 <|file_sep|>updated/requirements_dev.txt
# Database psycopg2-binary==2.7.7 pip==19.0.1 bumpversion==0.5.3 wheel==0.33.0 watchdog==0.9.0 flake8==3.7.5 tox==3.7.0 cryptography==2.5 Sphinx==1.8.4 coverage==4.5.2 pytest==4.2.0 pytest-runner==4.2 pytest-cov==2.6.1 # miscellaneous sphinx_rtd_theme==0.4.2 pre-commit==1.14.3
<|file_sep|>original/requirements_dev.txt # Database psycopg2-binary==2.7.7 pip==19.0.1 bumpversion==0.5.3 wheel==0.32.3 watchdog==0.9.0 flake8==3.7.5 tox==3.7.0 cryptography==2.5 Sphinx==1.8.4 coverage==4.5.2 pytest==4.2.0 pytest-runner==4.2 pytest-cov==2.6.1 # miscellaneous sphinx_rtd_theme==0.4.2 pre-commit==1.14.3 <|file_sep|>current/requirements_dev.txt # Database psycopg2-binary==2.7.7 pip==19.0.1 bumpversion==0.5.3 wheel==0.32.3 watchdog==0.9.0 flake8==3.7.5 tox==3.7.0 cryptography==2.5 Sphinx==1.8.4 coverage==4.5.2 pytest==4.2.0 pytest-runner==4.2 pytest-cov==2.6.1 # miscellaneous sphinx_rtd_theme==0.4.2 pre-commit==1.14.3 <|file_sep|>updated/requirements_dev.txt # Database psycopg2-binary==2.7.7 pip==19.0.1 bumpversion==0.5.3 wheel==0.33.0 watchdog==0.9.0 flake8==3.7.5 tox==3.7.0 cryptography==2.5 Sphinx==1.8.4 coverage==4.5.2 pytest==4.2.0 pytest-runner==4.2 pytest-cov==2.6.1 # miscellaneous sphinx_rtd_theme==0.4.2 pre-commit==1.14.3
0f41d6a1639841697d16fe0ee84664f8c41b3a78
requirements_dev.txt
requirements_dev.txt
Text
<|file_sep|>original/LICENSE.md The MIT License (MIT) Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. <|file_sep|>current/LICENSE.md The MIT License (MIT) Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. <|file_sep|>updated/LICENSE.md
The MIT License (MIT) Copyright (c) 2016 Patrick M. Jensen, patmjen@gmail.com Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
<|file_sep|>original/LICENSE.md The MIT License (MIT) Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. <|file_sep|>current/LICENSE.md The MIT License (MIT) Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. <|file_sep|>updated/LICENSE.md The MIT License (MIT) Copyright (c) 2016 Patrick M. Jensen, patmjen@gmail.com Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
148f46220e7c3028936032307706d4a7ddc85330
LICENSE.md
LICENSE.md
Markdown
<|file_sep|>original/app/widgets/ChatOmemo/locales.ini [omemo] fingerprints_title = Fingerprints fingerprints = OMEMO Fingerprints generating_bundle = "Generating the encryption keys, please wait…" generated_bundle = Encryption keys generated resolving_devices = Resolving OMEMO devices building_sessions = Building encryption sessions building_own_sessions = Building own encryption sessions encrypted = Encrypted message encrypted_disabled = Encryption disabled encrypted_loading = Send a message to build the encrypted session enable_contact = Enabling OMEMO for the contact disable_contact = Disabling OMEMO for the contact last_activity = Last activity sessions_built = %s sessions built <|file_sep|>current/app/widgets/ChatOmemo/locales.ini [omemo] fingerprints_title = Fingerprints fingerprints = OMEMO Fingerprints generating_bundle = "Generating the encryption keys, please wait…" generated_bundle = Encryption keys generated resolving_devices = Resolving OMEMO devices building_sessions = Building encryption sessions building_own_sessions = Building own encryption sessions encrypted = Encrypted message encrypted_disabled = Encryption disabled encrypted_loading = Send a message to build the encrypted session enable_contact = Enabling OMEMO for the contact disable_contact = Disabling OMEMO for the contact last_activity = Last activity sessions_built = %s sessions built <|file_sep|>updated/app/widgets/ChatOmemo/locales.ini
[omemo] fingerprints_title = Fingerprints fingerprints = OMEMO Fingerprints generating_bundle = "Generating the encryption keys, please wait…" generated_bundle = Encryption keys generated resolving_devices = Resolving OMEMO devices building_sessions = Building encryption sessions building_own_sessions = Building own encryption sessions encrypted = Encrypted message encrypted_disabled = Encryption disabled encrypted_loading = Send a message to build the encrypted session enable_contact = Enabling encryption for the contact disable_contact = Disabling encryption for the contact last_activity = Last activity sessions_built = %s sessions built
<|file_sep|>original/app/widgets/ChatOmemo/locales.ini [omemo] fingerprints_title = Fingerprints fingerprints = OMEMO Fingerprints generating_bundle = "Generating the encryption keys, please wait…" generated_bundle = Encryption keys generated resolving_devices = Resolving OMEMO devices building_sessions = Building encryption sessions building_own_sessions = Building own encryption sessions encrypted = Encrypted message encrypted_disabled = Encryption disabled encrypted_loading = Send a message to build the encrypted session enable_contact = Enabling OMEMO for the contact disable_contact = Disabling OMEMO for the contact last_activity = Last activity sessions_built = %s sessions built <|file_sep|>current/app/widgets/ChatOmemo/locales.ini [omemo] fingerprints_title = Fingerprints fingerprints = OMEMO Fingerprints generating_bundle = "Generating the encryption keys, please wait…" generated_bundle = Encryption keys generated resolving_devices = Resolving OMEMO devices building_sessions = Building encryption sessions building_own_sessions = Building own encryption sessions encrypted = Encrypted message encrypted_disabled = Encryption disabled encrypted_loading = Send a message to build the encrypted session enable_contact = Enabling OMEMO for the contact disable_contact = Disabling OMEMO for the contact last_activity = Last activity sessions_built = %s sessions built <|file_sep|>updated/app/widgets/ChatOmemo/locales.ini [omemo] fingerprints_title = Fingerprints fingerprints = OMEMO Fingerprints generating_bundle = "Generating the encryption keys, please wait…" generated_bundle = Encryption keys generated resolving_devices = Resolving OMEMO devices building_sessions = Building encryption sessions building_own_sessions = Building own encryption sessions encrypted = Encrypted message encrypted_disabled = Encryption disabled encrypted_loading = Send a message to build the encrypted session enable_contact = Enabling encryption for the contact disable_contact = Disabling encryption for the contact last_activity = Last activity sessions_built = %s sessions built
3a6212ea81b8689041e898df1fa458c6e1c7e806
app/widgets/ChatOmemo/locales.ini
app/widgets/ChatOmemo/locales.ini
INI
<|file_sep|>original/app/jobs/geocode_firm_job.rb require 'geocoder' class GeocodeFirmJob < ActiveJob::Base def perform(firm) coordinates = Geocoder.coordinates(firm.full_street_address) coordinates ? stat(:success) : stat(:failed) firm.geocode!(coordinates) end private def stat(key) Stats.increment("radsignup.geocode_firm.#{key}") end end <|file_sep|>current/app/jobs/geocode_firm_job.rb require 'geocoder' class GeocodeFirmJob < ActiveJob::Base def perform(firm) coordinates = Geocoder.coordinates(firm.full_street_address) coordinates ? stat(:success) : stat(:failed) firm.geocode!(coordinates) end private def stat(key) Stats.increment("radsignup.geocode_firm.#{key}") end end <|file_sep|>updated/app/jobs/geocode_firm_job.rb
require 'geocoder' class GeocodeFirmJob < ActiveJob::Base def perform(firm) coordinates = Geocoder.coordinates(firm.full_street_address) coordinates ? stat(:success) : stat(:failed) firm.geocode!(coordinates) end private def stat(key) Stats.increment("radsignup.geocode.firm.#{key}") end end
<|file_sep|>original/app/jobs/geocode_firm_job.rb require 'geocoder' class GeocodeFirmJob < ActiveJob::Base def perform(firm) coordinates = Geocoder.coordinates(firm.full_street_address) coordinates ? stat(:success) : stat(:failed) firm.geocode!(coordinates) end private def stat(key) Stats.increment("radsignup.geocode_firm.#{key}") end end <|file_sep|>current/app/jobs/geocode_firm_job.rb require 'geocoder' class GeocodeFirmJob < ActiveJob::Base def perform(firm) coordinates = Geocoder.coordinates(firm.full_street_address) coordinates ? stat(:success) : stat(:failed) firm.geocode!(coordinates) end private def stat(key) Stats.increment("radsignup.geocode_firm.#{key}") end end <|file_sep|>updated/app/jobs/geocode_firm_job.rb require 'geocoder' class GeocodeFirmJob < ActiveJob::Base def perform(firm) coordinates = Geocoder.coordinates(firm.full_street_address) coordinates ? stat(:success) : stat(:failed) firm.geocode!(coordinates) end private def stat(key) Stats.increment("radsignup.geocode.firm.#{key}") end end
3d03fc9be9ab1e2e661ca56ef87bf4f10b2fe993
app/jobs/geocode_firm_job.rb
app/jobs/geocode_firm_job.rb
Ruby
<|file_sep|>original/2d/skeleton/README.md # Skeleton2D Demo This demo shows how to create a rigged and animated character in 2D using Godot's Skeleton2D node. There are several movement-related animations and there is a simple character controller that controls the animations. Language: GDScript Renderer: GLES 2 ## Licenses GBot character Copyright &copy; circa 2020 Andreas Esau, MIT License. Initial rigging and animating Copyright &copy; 2020 RustyStriker, MIT License. ## Screenshots ![Screenshot](screenshots/screenshot.png) <|file_sep|>current/2d/skeleton/README.md # Skeleton2D Demo This demo shows how to create a rigged and animated character in 2D using Godot's Skeleton2D node. There are several movement-related animations and there is a simple character controller that controls the animations. Language: GDScript Renderer: GLES 2 ## Licenses GBot character Copyright &copy; circa 2020 Andreas Esau, MIT License. Initial rigging and animating Copyright &copy; 2020 RustyStriker, MIT License. ## Screenshots ![Screenshot](screenshots/screenshot.png) <|file_sep|>updated/2d/skeleton/README.md
# Skeleton2D Demo This demo shows how to create a rigged and animated character in 2D using Godot's Skeleton2D node. There are several movement-related animations and there is a simple character controller that controls the animations. Language: GDScript Renderer: GLES 2 Check out this demo on the asset library: https://godotengine.org/asset-library/asset/1027 ## Licenses GBot character Copyright &copy; circa 2020 Andreas Esau, MIT License. Initial rigging and animating Copyright &copy; 2020 RustyStriker, MIT License. ## Screenshots ![Screenshot](screenshots/screenshot.png)
<|file_sep|>original/2d/skeleton/README.md # Skeleton2D Demo This demo shows how to create a rigged and animated character in 2D using Godot's Skeleton2D node. There are several movement-related animations and there is a simple character controller that controls the animations. Language: GDScript Renderer: GLES 2 ## Licenses GBot character Copyright &copy; circa 2020 Andreas Esau, MIT License. Initial rigging and animating Copyright &copy; 2020 RustyStriker, MIT License. ## Screenshots ![Screenshot](screenshots/screenshot.png) <|file_sep|>current/2d/skeleton/README.md # Skeleton2D Demo This demo shows how to create a rigged and animated character in 2D using Godot's Skeleton2D node. There are several movement-related animations and there is a simple character controller that controls the animations. Language: GDScript Renderer: GLES 2 ## Licenses GBot character Copyright &copy; circa 2020 Andreas Esau, MIT License. Initial rigging and animating Copyright &copy; 2020 RustyStriker, MIT License. ## Screenshots ![Screenshot](screenshots/screenshot.png) <|file_sep|>updated/2d/skeleton/README.md # Skeleton2D Demo This demo shows how to create a rigged and animated character in 2D using Godot's Skeleton2D node. There are several movement-related animations and there is a simple character controller that controls the animations. Language: GDScript Renderer: GLES 2 Check out this demo on the asset library: https://godotengine.org/asset-library/asset/1027 ## Licenses GBot character Copyright &copy; circa 2020 Andreas Esau, MIT License. Initial rigging and animating Copyright &copy; 2020 RustyStriker, MIT License. ## Screenshots ![Screenshot](screenshots/screenshot.png)
9ea18370d3ba71d2bad7b944eb521dc309b3f513
2d/skeleton/README.md
2d/skeleton/README.md
Markdown
<|file_sep|>original/requirements/prod.txt beautifulsoup4==4.8.2 eventlet==0.25.1 gunicorn==20.0.4 # CSS/JS Assets flask_assets==2.0 yuicompressor==2.4.8 # Sockets flask_socketio==4.2.1 python_socketio==4.4.0 python_engineio==3.11.2 # Celery celery==4.4.1 redis==3.4.1 # Octave interaction scipy==1.4.1 numpy==1.18.1 octave-kernel==0.31.0 <|file_sep|>current/requirements/prod.txt beautifulsoup4==4.8.2 eventlet==0.25.1 gunicorn==20.0.4 # CSS/JS Assets flask_assets==2.0 yuicompressor==2.4.8 # Sockets flask_socketio==4.2.1 python_socketio==4.4.0 python_engineio==3.11.2 # Celery celery==4.4.1 redis==3.4.1 # Octave interaction scipy==1.4.1 numpy==1.18.1 octave-kernel==0.31.0 <|file_sep|>updated/requirements/prod.txt
beautifulsoup4==4.8.2 eventlet==0.25.1 gunicorn==20.0.4 # CSS/JS Assets flask_assets==2.0 yuicompressor==2.4.8 # Sockets flask_socketio==4.2.1 python_socketio==4.5.0 python_engineio==3.11.2 # Celery celery==4.4.1 redis==3.4.1 # Octave interaction scipy==1.4.1 numpy==1.18.1 octave-kernel==0.31.0
<|file_sep|>original/requirements/prod.txt beautifulsoup4==4.8.2 eventlet==0.25.1 gunicorn==20.0.4 # CSS/JS Assets flask_assets==2.0 yuicompressor==2.4.8 # Sockets flask_socketio==4.2.1 python_socketio==4.4.0 python_engineio==3.11.2 # Celery celery==4.4.1 redis==3.4.1 # Octave interaction scipy==1.4.1 numpy==1.18.1 octave-kernel==0.31.0 <|file_sep|>current/requirements/prod.txt beautifulsoup4==4.8.2 eventlet==0.25.1 gunicorn==20.0.4 # CSS/JS Assets flask_assets==2.0 yuicompressor==2.4.8 # Sockets flask_socketio==4.2.1 python_socketio==4.4.0 python_engineio==3.11.2 # Celery celery==4.4.1 redis==3.4.1 # Octave interaction scipy==1.4.1 numpy==1.18.1 octave-kernel==0.31.0 <|file_sep|>updated/requirements/prod.txt beautifulsoup4==4.8.2 eventlet==0.25.1 gunicorn==20.0.4 # CSS/JS Assets flask_assets==2.0 yuicompressor==2.4.8 # Sockets flask_socketio==4.2.1 python_socketio==4.5.0 python_engineio==3.11.2 # Celery celery==4.4.1 redis==3.4.1 # Octave interaction scipy==1.4.1 numpy==1.18.1 octave-kernel==0.31.0
3c0e8bc67d8226e121f18ff917cde81ed46e8a94
requirements/prod.txt
requirements/prod.txt
Text
<|file_sep|>original/package.json "main": "index.js", "scripts": { "test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;" }, "author": "Edward Faulkner <ef@alum.mit.edu>", "license": "MIT", "repository": "https://github.com/ef4/ember-code-snippet", "dependencies": { "highlight.js": "^8.1.0", "broccoli-static-compiler": "^0.1.4", "broccoli-merge-trees": "^0.1.4", "broccoli-browserify": "^0.1.0", "broccoli-spelunk": "^0.1.2", "es6-promise": "^1.0.0", "broccoli-writer": "^0.1.1", "glob": "^4.0.4" }, "keywords": [ "ember-addon" ] } <|file_sep|>current/package.json "main": "index.js", "scripts": { "test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;" }, "author": "Edward Faulkner <ef@alum.mit.edu>", "license": "MIT", "repository": "https://github.com/ef4/ember-code-snippet", "dependencies": { "highlight.js": "^8.1.0", "broccoli-static-compiler": "^0.1.4", "broccoli-merge-trees": "^0.1.4", "broccoli-browserify": "^0.1.0", "broccoli-spelunk": "^0.1.2", "es6-promise": "^1.0.0", "broccoli-writer": "^0.1.1", "glob": "^4.0.4" }, "keywords": [ "ember-addon" ] } <|file_sep|>updated/package.json
"main": "index.js", "scripts": { "test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;" }, "author": "Edward Faulkner <ef@alum.mit.edu>", "license": "MIT", "repository": "https://github.com/ef4/ember-code-snippet", "dependencies": { "highlight.js": "^8.1.0", "broccoli-static-compiler": "^0.1.4", "broccoli-merge-trees": "^0.1.4", "broccoli-browserify": "^0.1.0", "broccoli-flatiron": "^0.0.0", "es6-promise": "^1.0.0", "broccoli-writer": "^0.1.1", "glob": "^4.0.4" }, "keywords": [ "ember-addon" ] }
<|file_sep|>original/package.json "main": "index.js", "scripts": { "test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;" }, "author": "Edward Faulkner <ef@alum.mit.edu>", "license": "MIT", "repository": "https://github.com/ef4/ember-code-snippet", "dependencies": { "highlight.js": "^8.1.0", "broccoli-static-compiler": "^0.1.4", "broccoli-merge-trees": "^0.1.4", "broccoli-browserify": "^0.1.0", "broccoli-spelunk": "^0.1.2", "es6-promise": "^1.0.0", "broccoli-writer": "^0.1.1", "glob": "^4.0.4" }, "keywords": [ "ember-addon" ] } <|file_sep|>current/package.json "main": "index.js", "scripts": { "test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;" }, "author": "Edward Faulkner <ef@alum.mit.edu>", "license": "MIT", "repository": "https://github.com/ef4/ember-code-snippet", "dependencies": { "highlight.js": "^8.1.0", "broccoli-static-compiler": "^0.1.4", "broccoli-merge-trees": "^0.1.4", "broccoli-browserify": "^0.1.0", "broccoli-spelunk": "^0.1.2", "es6-promise": "^1.0.0", "broccoli-writer": "^0.1.1", "glob": "^4.0.4" }, "keywords": [ "ember-addon" ] } <|file_sep|>updated/package.json "main": "index.js", "scripts": { "test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;" }, "author": "Edward Faulkner <ef@alum.mit.edu>", "license": "MIT", "repository": "https://github.com/ef4/ember-code-snippet", "dependencies": { "highlight.js": "^8.1.0", "broccoli-static-compiler": "^0.1.4", "broccoli-merge-trees": "^0.1.4", "broccoli-browserify": "^0.1.0", "broccoli-flatiron": "^0.0.0", "es6-promise": "^1.0.0", "broccoli-writer": "^0.1.1", "glob": "^4.0.4" }, "keywords": [ "ember-addon" ] }
91f8e4f0f4a5d06e4a6fd9b622966df633c2e34b
package.json
package.json
JSON
<|file_sep|>lib/alchemy/logger.rb.diff original: # Logs a warning to the Rails standard logger and adds some nicer formatting updated: # Logs a debug message to the Rails standard logger and adds some nicer formatting <|file_sep|>original/lib/alchemy/logger.rb module Alchemy module Logger # Logs a warning to the Rails standard logger and adds some nicer formatting def self.warn(message, caller_string) Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n) return nil end def log_warning(message) Alchemy::Logger.warn(message, caller.first) end end end <|file_sep|>current/lib/alchemy/logger.rb module Alchemy module Logger # Logs a debug message to the Rails standard logger and adds some nicer formatting def self.warn(message, caller_string) Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n) return nil end def log_warning(message) Alchemy::Logger.warn(message, caller.first) end end end <|file_sep|>updated/lib/alchemy/logger.rb
module Alchemy module Logger # Logs a debug message to the Rails standard logger and adds some nicer formatting def self.warn(message, caller_string) Rails.logger.debug %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n) return nil end def log_warning(message) Alchemy::Logger.warn(message, caller.first) end end end
<|file_sep|>lib/alchemy/logger.rb.diff original: # Logs a warning to the Rails standard logger and adds some nicer formatting updated: # Logs a debug message to the Rails standard logger and adds some nicer formatting <|file_sep|>original/lib/alchemy/logger.rb module Alchemy module Logger # Logs a warning to the Rails standard logger and adds some nicer formatting def self.warn(message, caller_string) Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n) return nil end def log_warning(message) Alchemy::Logger.warn(message, caller.first) end end end <|file_sep|>current/lib/alchemy/logger.rb module Alchemy module Logger # Logs a debug message to the Rails standard logger and adds some nicer formatting def self.warn(message, caller_string) Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n) return nil end def log_warning(message) Alchemy::Logger.warn(message, caller.first) end end end <|file_sep|>updated/lib/alchemy/logger.rb module Alchemy module Logger # Logs a debug message to the Rails standard logger and adds some nicer formatting def self.warn(message, caller_string) Rails.logger.debug %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n) return nil end def log_warning(message) Alchemy::Logger.warn(message, caller.first) end end end
07f8826eddc964d46e3476c279f7819ec9e1cfd1
lib/alchemy/logger.rb
lib/alchemy/logger.rb
Ruby
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: updated: use Psy\Input\CodeArgument; <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: updated: use Symfony\Component\Console\Input\InputArgument; <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: use Symfony\Component\Console\Input\InputArgument; updated: <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: new InputArgument('target', InputArgument::REQUIRED, 'A target object or primitive to profile.', null), updated: new CodeArgument('code', InputArgument::REQUIRED, 'Code to execute.'), <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: $target = $input->getArgument('target'); $start = microtime(true); updated: $code = $input->getArgument('code'); <|file_sep|>original/src/Psy/Command/TimeitCommand.php } /** * {@inheritdoc} */ protected function execute(InputInterface $input, OutputInterface $output) { $target = $input->getArgument('target'); $start = microtime(true); /** @var Shell $shell */ $shell = $this->getApplication(); $sh = new Shell(new Configuration()); $sh->setOutput($output); $sh->setScopeVariables($shell->getScopeVariables()); $sh->execute($target); $end = microtime(true); $output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start)); } <|file_sep|>current/src/Psy/Command/TimeitCommand.php ); } /** * {@inheritdoc} */ protected function execute(InputInterface $input, OutputInterface $output) { $code = $input->getArgument('code'); /** @var Shell $shell */ $shell = $this->getApplication(); $sh = new Shell(new Configuration()); $sh->setOutput($output); $sh->setScopeVariables($shell->getScopeVariables()); $end = microtime(true); $output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start)); } } <|file_sep|>updated/src/Psy/Command/TimeitCommand.php
/** * {@inheritdoc} */ protected function execute(InputInterface $input, OutputInterface $output) { $code = $input->getArgument('code'); /** @var Shell $shell */ $shell = $this->getApplication(); $sh = new Shell(new Configuration()); $sh->setOutput($output); $sh->setScopeVariables($shell->getScopeVariables()); $start = microtime(true); $sh->execute($code); $end = microtime(true); $output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start)); } }
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: updated: use Psy\Input\CodeArgument; <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: updated: use Symfony\Component\Console\Input\InputArgument; <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: use Symfony\Component\Console\Input\InputArgument; updated: <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: new InputArgument('target', InputArgument::REQUIRED, 'A target object or primitive to profile.', null), updated: new CodeArgument('code', InputArgument::REQUIRED, 'Code to execute.'), <|file_sep|>src/Psy/Command/TimeitCommand.php.diff original: $target = $input->getArgument('target'); $start = microtime(true); updated: $code = $input->getArgument('code'); <|file_sep|>original/src/Psy/Command/TimeitCommand.php } /** * {@inheritdoc} */ protected function execute(InputInterface $input, OutputInterface $output) { $target = $input->getArgument('target'); $start = microtime(true); /** @var Shell $shell */ $shell = $this->getApplication(); $sh = new Shell(new Configuration()); $sh->setOutput($output); $sh->setScopeVariables($shell->getScopeVariables()); $sh->execute($target); $end = microtime(true); $output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start)); } <|file_sep|>current/src/Psy/Command/TimeitCommand.php ); } /** * {@inheritdoc} */ protected function execute(InputInterface $input, OutputInterface $output) { $code = $input->getArgument('code'); /** @var Shell $shell */ $shell = $this->getApplication(); $sh = new Shell(new Configuration()); $sh->setOutput($output); $sh->setScopeVariables($shell->getScopeVariables()); $end = microtime(true); $output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start)); } } <|file_sep|>updated/src/Psy/Command/TimeitCommand.php /** * {@inheritdoc} */ protected function execute(InputInterface $input, OutputInterface $output) { $code = $input->getArgument('code'); /** @var Shell $shell */ $shell = $this->getApplication(); $sh = new Shell(new Configuration()); $sh->setOutput($output); $sh->setScopeVariables($shell->getScopeVariables()); $start = microtime(true); $sh->execute($code); $end = microtime(true); $output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start)); } }
eee4305f71d53dfaeed6b0042926e74f8530bece
src/Psy/Command/TimeitCommand.php
src/Psy/Command/TimeitCommand.php
PHP
<|file_sep|>original/.travis.yml - python: 2.7 env: TOXENV=py27 - python: 3.4 env: TOXENV=py34 - python: 3.5 env: TOXENV=py35 - python: 3.6 env: TOXENV=py36 - python: 3.7 env: TOXENV=py37 - python: 3.8 env: TOXENV=py38 - python: 3.5 env: TOXENV=flake - python: 3.5 env: TOXENV=cover install: - pip install tox script: - tox after_script: <|file_sep|>current/.travis.yml - python: 2.7 env: TOXENV=py27 - python: 3.4 env: TOXENV=py34 - python: 3.5 env: TOXENV=py35 - python: 3.6 env: TOXENV=py36 - python: 3.7 env: TOXENV=py37 - python: 3.8 env: TOXENV=py38 - python: 3.5 env: TOXENV=flake - python: 3.5 env: TOXENV=cover install: - pip install tox script: - tox after_script: <|file_sep|>updated/.travis.yml
- python: 2.7 env: TOXENV=py27 - python: 3.4 env: TOXENV=py34 - python: 3.5 env: TOXENV=py35 - python: 3.6 env: TOXENV=py36 - python: 3.7 env: TOXENV=py37 - python: 3.8-dev env: TOXENV=py38 - python: 3.5 env: TOXENV=flake - python: 3.5 env: TOXENV=cover install: - pip install tox script: - tox after_script:
<|file_sep|>original/.travis.yml - python: 2.7 env: TOXENV=py27 - python: 3.4 env: TOXENV=py34 - python: 3.5 env: TOXENV=py35 - python: 3.6 env: TOXENV=py36 - python: 3.7 env: TOXENV=py37 - python: 3.8 env: TOXENV=py38 - python: 3.5 env: TOXENV=flake - python: 3.5 env: TOXENV=cover install: - pip install tox script: - tox after_script: <|file_sep|>current/.travis.yml - python: 2.7 env: TOXENV=py27 - python: 3.4 env: TOXENV=py34 - python: 3.5 env: TOXENV=py35 - python: 3.6 env: TOXENV=py36 - python: 3.7 env: TOXENV=py37 - python: 3.8 env: TOXENV=py38 - python: 3.5 env: TOXENV=flake - python: 3.5 env: TOXENV=cover install: - pip install tox script: - tox after_script: <|file_sep|>updated/.travis.yml - python: 2.7 env: TOXENV=py27 - python: 3.4 env: TOXENV=py34 - python: 3.5 env: TOXENV=py35 - python: 3.6 env: TOXENV=py36 - python: 3.7 env: TOXENV=py37 - python: 3.8-dev env: TOXENV=py38 - python: 3.5 env: TOXENV=flake - python: 3.5 env: TOXENV=cover install: - pip install tox script: - tox after_script:
269134b873751b0fe68e5442f04e0ef36656568c
.travis.yml
.travis.yml
YAML
<|file_sep|>original/recipes/trigger_recipe_roller.txt No-op file. Edit this to kick recipes. This is a beginning of a story in this silly file. Once upon a time, a budding web browser dev team needed a CI system. All they had was one poor machine under a desk, and its name was Batty, the Build and Test Yeti. As the CI needs of the browser grew, Batty, the Build and Test Yeti, got a new friend: The End. <|file_sep|>current/recipes/trigger_recipe_roller.txt No-op file. Edit this to kick recipes. This is a beginning of a story in this silly file. Once upon a time, a budding web browser dev team needed a CI system. All they had was one poor machine under a desk, and its name was Batty, the Build and Test Yeti. As the CI needs of the browser grew, Batty, the Build and Test Yeti, got a new friend: The End. <|file_sep|>updated/recipes/trigger_recipe_roller.txt
No-op file. Edit this to kick recipes. This is a beginning of a story in this silly file. Once upon a time, a budding web browser dev team needed a CI system. All they had was one poor machine under a desk, and its name was Batty, the Build and Test Yeti. As the CI needs of the browser grew, Batty, the Build and Test Yeti, got a new friend: The End!
<|file_sep|>original/recipes/trigger_recipe_roller.txt No-op file. Edit this to kick recipes. This is a beginning of a story in this silly file. Once upon a time, a budding web browser dev team needed a CI system. All they had was one poor machine under a desk, and its name was Batty, the Build and Test Yeti. As the CI needs of the browser grew, Batty, the Build and Test Yeti, got a new friend: The End. <|file_sep|>current/recipes/trigger_recipe_roller.txt No-op file. Edit this to kick recipes. This is a beginning of a story in this silly file. Once upon a time, a budding web browser dev team needed a CI system. All they had was one poor machine under a desk, and its name was Batty, the Build and Test Yeti. As the CI needs of the browser grew, Batty, the Build and Test Yeti, got a new friend: The End. <|file_sep|>updated/recipes/trigger_recipe_roller.txt No-op file. Edit this to kick recipes. This is a beginning of a story in this silly file. Once upon a time, a budding web browser dev team needed a CI system. All they had was one poor machine under a desk, and its name was Batty, the Build and Test Yeti. As the CI needs of the browser grew, Batty, the Build and Test Yeti, got a new friend: The End!
b63c4663d478bcf8d4fd163a52dc42eb85861ea7
recipes/trigger_recipe_roller.txt
recipes/trigger_recipe_roller.txt
Text
<|file_sep|>teknologr/members/static/js/applicant.js.diff original: $('#unknown_degree input').val(''); updated: $('#id_degree_programme').val(''); <|file_sep|>original/teknologr/members/static/js/applicant.js $(document).ready(function() { $('#id_degree_programme_options').change(function() { console.log('hello'); if (this.value === 'extra') { $('#unknown_degree').show(); $('#unknown_degree input').val(''); } else { $('#unknown_degree').hide(); $('#unknown_degree input').val(this.value); } }); }); <|file_sep|>current/teknologr/members/static/js/applicant.js $(document).ready(function() { $('#id_degree_programme_options').change(function() { console.log('hello'); if (this.value === 'extra') { $('#unknown_degree').show(); $('#id_degree_programme').val(''); } else { $('#unknown_degree').hide(); $('#unknown_degree input').val(this.value); } }); }); <|file_sep|>updated/teknologr/members/static/js/applicant.js
$(document).ready(function() { $('#id_degree_programme_options').change(function() { console.log('hello'); if (this.value === 'extra') { $('#unknown_degree').show(); $('#id_degree_programme').val(''); } else { $('#unknown_degree').hide(); $('#id_degree_programme').val(this.value); } }); });
<|file_sep|>teknologr/members/static/js/applicant.js.diff original: $('#unknown_degree input').val(''); updated: $('#id_degree_programme').val(''); <|file_sep|>original/teknologr/members/static/js/applicant.js $(document).ready(function() { $('#id_degree_programme_options').change(function() { console.log('hello'); if (this.value === 'extra') { $('#unknown_degree').show(); $('#unknown_degree input').val(''); } else { $('#unknown_degree').hide(); $('#unknown_degree input').val(this.value); } }); }); <|file_sep|>current/teknologr/members/static/js/applicant.js $(document).ready(function() { $('#id_degree_programme_options').change(function() { console.log('hello'); if (this.value === 'extra') { $('#unknown_degree').show(); $('#id_degree_programme').val(''); } else { $('#unknown_degree').hide(); $('#unknown_degree input').val(this.value); } }); }); <|file_sep|>updated/teknologr/members/static/js/applicant.js $(document).ready(function() { $('#id_degree_programme_options').change(function() { console.log('hello'); if (this.value === 'extra') { $('#unknown_degree').show(); $('#id_degree_programme').val(''); } else { $('#unknown_degree').hide(); $('#id_degree_programme').val(this.value); } }); });
d6ce1424c5d99ba80ad0f13532d20140ae195e60
teknologr/members/static/js/applicant.js
teknologr/members/static/js/applicant.js
JavaScript
<|file_sep|>original/LICENSE.txt <|file_sep|>current/LICENSE.txt <|file_sep|>updated/LICENSE.txt
Copyright (c) 2013, Sergey Vinokurov All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the <organization> nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL <COPYRIGHT HOLDER> BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
<|file_sep|>original/LICENSE.txt <|file_sep|>current/LICENSE.txt <|file_sep|>updated/LICENSE.txt Copyright (c) 2013, Sergey Vinokurov All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the <organization> nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL <COPYRIGHT HOLDER> BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
8654eba3046dcc0eab7ce43497e58df05f719480
LICENSE.txt
LICENSE.txt
Text
<|file_sep|>src/main/webapp/src/app/app.module.ts.diff original: import {TestPageComponent} from './test-page/test-page.component'; updated: <|file_sep|>original/src/main/webapp/src/app/app.module.ts import {BrowserModule} from '@angular/platform-browser'; import {NgModule} from '@angular/core'; import {AppRoutingModule} from './app-routing.module'; import {AppComponent} from './app.component'; import {MaterialComponents} from './material-components'; import {TestPageComponent} from './test-page/test-page.component'; import {BrowserAnimationsModule} from '@angular/platform-browser/animations'; @NgModule({ declarations: [AppComponent, TestPageComponent], imports: [ MaterialComponents, BrowserModule, AppRoutingModule, BrowserAnimationsModule, ], providers: [], bootstrap: [AppComponent], }) <|file_sep|>current/src/main/webapp/src/app/app.module.ts import {BrowserModule} from '@angular/platform-browser'; import {NgModule} from '@angular/core'; import {AppRoutingModule} from './app-routing.module'; import {AppComponent} from './app.component'; import {MaterialComponents} from './material-components'; import {BrowserAnimationsModule} from '@angular/platform-browser/animations'; @NgModule({ declarations: [AppComponent, TestPageComponent], imports: [ MaterialComponents, BrowserModule, AppRoutingModule, BrowserAnimationsModule, ], providers: [], bootstrap: [AppComponent], }) export class AppModule {} <|file_sep|>updated/src/main/webapp/src/app/app.module.ts
import {BrowserModule} from '@angular/platform-browser'; import {NgModule} from '@angular/core'; import {AppRoutingModule} from './app-routing.module'; import {AppComponent} from './app.component'; import {MaterialComponents} from './material-components'; import {BrowserAnimationsModule} from '@angular/platform-browser/animations'; @NgModule({ declarations: [AppComponent], imports: [ MaterialComponents, BrowserModule, AppRoutingModule, BrowserAnimationsModule, ], providers: [], bootstrap: [AppComponent], }) export class AppModule {}
<|file_sep|>src/main/webapp/src/app/app.module.ts.diff original: import {TestPageComponent} from './test-page/test-page.component'; updated: <|file_sep|>original/src/main/webapp/src/app/app.module.ts import {BrowserModule} from '@angular/platform-browser'; import {NgModule} from '@angular/core'; import {AppRoutingModule} from './app-routing.module'; import {AppComponent} from './app.component'; import {MaterialComponents} from './material-components'; import {TestPageComponent} from './test-page/test-page.component'; import {BrowserAnimationsModule} from '@angular/platform-browser/animations'; @NgModule({ declarations: [AppComponent, TestPageComponent], imports: [ MaterialComponents, BrowserModule, AppRoutingModule, BrowserAnimationsModule, ], providers: [], bootstrap: [AppComponent], }) <|file_sep|>current/src/main/webapp/src/app/app.module.ts import {BrowserModule} from '@angular/platform-browser'; import {NgModule} from '@angular/core'; import {AppRoutingModule} from './app-routing.module'; import {AppComponent} from './app.component'; import {MaterialComponents} from './material-components'; import {BrowserAnimationsModule} from '@angular/platform-browser/animations'; @NgModule({ declarations: [AppComponent, TestPageComponent], imports: [ MaterialComponents, BrowserModule, AppRoutingModule, BrowserAnimationsModule, ], providers: [], bootstrap: [AppComponent], }) export class AppModule {} <|file_sep|>updated/src/main/webapp/src/app/app.module.ts import {BrowserModule} from '@angular/platform-browser'; import {NgModule} from '@angular/core'; import {AppRoutingModule} from './app-routing.module'; import {AppComponent} from './app.component'; import {MaterialComponents} from './material-components'; import {BrowserAnimationsModule} from '@angular/platform-browser/animations'; @NgModule({ declarations: [AppComponent], imports: [ MaterialComponents, BrowserModule, AppRoutingModule, BrowserAnimationsModule, ], providers: [], bootstrap: [AppComponent], }) export class AppModule {}
02e010d7e85345d26632897972bb735134bf225d
src/main/webapp/src/app/app.module.ts
src/main/webapp/src/app/app.module.ts
TypeScript
<|file_sep|>original/dist_chrome/content-script.js } }); port.start(); } // let ember-debug know that content script has executed document.documentElement.dataset.emberExtension = 1; // Allow older versions of Ember (< 1.4) to detect the extension. if (document.body) { document.body.dataset.emberExtension = 1; } // clear a possible previous Ember icon chrome.extension.sendMessage({ type: 'resetEmberIcon' }); // inject JS into the page to check for an app on domready var script = document.createElement('script'); script.type = "text/javascript"; script.src = chrome.extension.getURL("in-page-script.js"); <|file_sep|>current/dist_chrome/content-script.js } }); port.start(); } // let ember-debug know that content script has executed document.documentElement.dataset.emberExtension = 1; // Allow older versions of Ember (< 1.4) to detect the extension. if (document.body) { document.body.dataset.emberExtension = 1; } // clear a possible previous Ember icon chrome.extension.sendMessage({ type: 'resetEmberIcon' }); // inject JS into the page to check for an app on domready var script = document.createElement('script'); script.type = "text/javascript"; script.src = chrome.extension.getURL("in-page-script.js"); <|file_sep|>updated/dist_chrome/content-script.js
port.postMessage(message); } }); port.start(); } // let ember-debug know that content script has executed document.documentElement.dataset.emberExtension = 1; // clear a possible previous Ember icon chrome.extension.sendMessage({ type: 'resetEmberIcon' }); // inject JS into the page to check for an app on domready var script = document.createElement('script'); script.type = "text/javascript"; script.src = chrome.extension.getURL("in-page-script.js"); if (document.body) document.body.appendChild(script); }());
<|file_sep|>original/dist_chrome/content-script.js } }); port.start(); } // let ember-debug know that content script has executed document.documentElement.dataset.emberExtension = 1; // Allow older versions of Ember (< 1.4) to detect the extension. if (document.body) { document.body.dataset.emberExtension = 1; } // clear a possible previous Ember icon chrome.extension.sendMessage({ type: 'resetEmberIcon' }); // inject JS into the page to check for an app on domready var script = document.createElement('script'); script.type = "text/javascript"; script.src = chrome.extension.getURL("in-page-script.js"); <|file_sep|>current/dist_chrome/content-script.js } }); port.start(); } // let ember-debug know that content script has executed document.documentElement.dataset.emberExtension = 1; // Allow older versions of Ember (< 1.4) to detect the extension. if (document.body) { document.body.dataset.emberExtension = 1; } // clear a possible previous Ember icon chrome.extension.sendMessage({ type: 'resetEmberIcon' }); // inject JS into the page to check for an app on domready var script = document.createElement('script'); script.type = "text/javascript"; script.src = chrome.extension.getURL("in-page-script.js"); <|file_sep|>updated/dist_chrome/content-script.js port.postMessage(message); } }); port.start(); } // let ember-debug know that content script has executed document.documentElement.dataset.emberExtension = 1; // clear a possible previous Ember icon chrome.extension.sendMessage({ type: 'resetEmberIcon' }); // inject JS into the page to check for an app on domready var script = document.createElement('script'); script.type = "text/javascript"; script.src = chrome.extension.getURL("in-page-script.js"); if (document.body) document.body.appendChild(script); }());
ae530367b59848c2259bf50e85e087634dab9bb1
dist_chrome/content-script.js
dist_chrome/content-script.js
JavaScript
<|file_sep|>ElCapitan/brew-casks.sh.diff original: updated: brew cask install cyberduck <|file_sep|>original/ElCapitan/brew-casks.sh echo "Installing brew casks" brew cask install alfred brew cask install android-studio brew cask install appcleaner brew cask install atom brew cask install carbon-copy-cloner brew cask install ccmenu brew cask install dash brew cask install dropbox brew cask install firefox brew cask install flux brew cask install handbrake brew cask install Caskroom/versions/intellij-idea-ce brew cask install kindle brew cask install macdown brew cask install opera brew cask install skitch brew cask install steam brew cask install textwrangler brew cask install visual-studio-code brew cask install vlc <|file_sep|>current/ElCapitan/brew-casks.sh brew cask install alfred brew cask install android-studio brew cask install appcleaner brew cask install atom brew cask install carbon-copy-cloner brew cask install ccmenu brew cask install cyberduck brew cask install dash brew cask install dropbox brew cask install firefox brew cask install flux brew cask install handbrake brew cask install Caskroom/versions/intellij-idea-ce brew cask install kindle brew cask install macdown brew cask install opera brew cask install skitch brew cask install steam brew cask install textwrangler brew cask install visual-studio-code brew cask install vlc <|file_sep|>updated/ElCapitan/brew-casks.sh
brew cask install android-studio brew cask install appcleaner brew cask install atom brew cask install carbon-copy-cloner brew cask install ccmenu brew cask install cyberduck brew cask install dash brew cask install dropbox brew cask install firefox brew cask install flux brew cask install handbrake brew cask install Caskroom/versions/intellij-idea-ce brew cask install kindle brew cask install macdown brew cask install opera brew cask install skitch brew cask install steam brew cask install textwrangler brew cask install visual-studio-code brew cask install vlc brew cask install xscope
<|file_sep|>ElCapitan/brew-casks.sh.diff original: updated: brew cask install cyberduck <|file_sep|>original/ElCapitan/brew-casks.sh echo "Installing brew casks" brew cask install alfred brew cask install android-studio brew cask install appcleaner brew cask install atom brew cask install carbon-copy-cloner brew cask install ccmenu brew cask install dash brew cask install dropbox brew cask install firefox brew cask install flux brew cask install handbrake brew cask install Caskroom/versions/intellij-idea-ce brew cask install kindle brew cask install macdown brew cask install opera brew cask install skitch brew cask install steam brew cask install textwrangler brew cask install visual-studio-code brew cask install vlc <|file_sep|>current/ElCapitan/brew-casks.sh brew cask install alfred brew cask install android-studio brew cask install appcleaner brew cask install atom brew cask install carbon-copy-cloner brew cask install ccmenu brew cask install cyberduck brew cask install dash brew cask install dropbox brew cask install firefox brew cask install flux brew cask install handbrake brew cask install Caskroom/versions/intellij-idea-ce brew cask install kindle brew cask install macdown brew cask install opera brew cask install skitch brew cask install steam brew cask install textwrangler brew cask install visual-studio-code brew cask install vlc <|file_sep|>updated/ElCapitan/brew-casks.sh brew cask install android-studio brew cask install appcleaner brew cask install atom brew cask install carbon-copy-cloner brew cask install ccmenu brew cask install cyberduck brew cask install dash brew cask install dropbox brew cask install firefox brew cask install flux brew cask install handbrake brew cask install Caskroom/versions/intellij-idea-ce brew cask install kindle brew cask install macdown brew cask install opera brew cask install skitch brew cask install steam brew cask install textwrangler brew cask install visual-studio-code brew cask install vlc brew cask install xscope
c561feb49f4d60bd4dc189f2617e6101cd717824
ElCapitan/brew-casks.sh
ElCapitan/brew-casks.sh
Shell
<|file_sep|>original/thrifty-bunny.gemspec spec.email = ["bsiggelkow@me.com"] spec.summary = "RabbitMQ adapter for Apache Thrift" spec.description = "RabbitMQ adapter for Apache Thrift" spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny" spec.license = 'MIT' spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^(test|spec|features)/}) spec.require_paths = ["lib"] spec.add_development_dependency "bundler", "~> 1.7" spec.add_development_dependency "rake", "~> 10.0" spec.add_development_dependency 'rspec' spec.add_development_dependency 'pry' spec.add_dependency 'thrift' spec.add_dependency 'thin' spec.add_dependency 'bunny' spec.add_dependency 'uuidtools' end <|file_sep|>current/thrifty-bunny.gemspec spec.email = ["bsiggelkow@me.com"] spec.summary = "RabbitMQ adapter for Apache Thrift" spec.description = "RabbitMQ adapter for Apache Thrift" spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny" spec.license = 'MIT' spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^(test|spec|features)/}) spec.require_paths = ["lib"] spec.add_development_dependency "bundler", "~> 1.7" spec.add_development_dependency "rake", "~> 10.0" spec.add_development_dependency 'rspec' spec.add_development_dependency 'pry' spec.add_dependency 'thrift' spec.add_dependency 'thin' spec.add_dependency 'bunny' spec.add_dependency 'uuidtools' end <|file_sep|>updated/thrifty-bunny.gemspec
spec.email = ["bsiggelkow@me.com"] spec.summary = "RabbitMQ adapter for Apache Thrift" spec.description = "RabbitMQ adapter for Apache Thrift" spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny" spec.license = 'MIT' spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^(test|spec|features)/}) spec.require_paths = ["lib"] spec.add_development_dependency "bundler", "~> 1.7" spec.add_development_dependency "rake", "~> 10.0" spec.add_development_dependency 'rspec' spec.add_development_dependency 'pry' spec.add_dependency 'thrift' spec.add_dependency 'thin' spec.add_dependency 'bunny', '~> 1.6.3' spec.add_dependency 'uuidtools' end
<|file_sep|>original/thrifty-bunny.gemspec spec.email = ["bsiggelkow@me.com"] spec.summary = "RabbitMQ adapter for Apache Thrift" spec.description = "RabbitMQ adapter for Apache Thrift" spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny" spec.license = 'MIT' spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^(test|spec|features)/}) spec.require_paths = ["lib"] spec.add_development_dependency "bundler", "~> 1.7" spec.add_development_dependency "rake", "~> 10.0" spec.add_development_dependency 'rspec' spec.add_development_dependency 'pry' spec.add_dependency 'thrift' spec.add_dependency 'thin' spec.add_dependency 'bunny' spec.add_dependency 'uuidtools' end <|file_sep|>current/thrifty-bunny.gemspec spec.email = ["bsiggelkow@me.com"] spec.summary = "RabbitMQ adapter for Apache Thrift" spec.description = "RabbitMQ adapter for Apache Thrift" spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny" spec.license = 'MIT' spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^(test|spec|features)/}) spec.require_paths = ["lib"] spec.add_development_dependency "bundler", "~> 1.7" spec.add_development_dependency "rake", "~> 10.0" spec.add_development_dependency 'rspec' spec.add_development_dependency 'pry' spec.add_dependency 'thrift' spec.add_dependency 'thin' spec.add_dependency 'bunny' spec.add_dependency 'uuidtools' end <|file_sep|>updated/thrifty-bunny.gemspec spec.email = ["bsiggelkow@me.com"] spec.summary = "RabbitMQ adapter for Apache Thrift" spec.description = "RabbitMQ adapter for Apache Thrift" spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny" spec.license = 'MIT' spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^(test|spec|features)/}) spec.require_paths = ["lib"] spec.add_development_dependency "bundler", "~> 1.7" spec.add_development_dependency "rake", "~> 10.0" spec.add_development_dependency 'rspec' spec.add_development_dependency 'pry' spec.add_dependency 'thrift' spec.add_dependency 'thin' spec.add_dependency 'bunny', '~> 1.6.3' spec.add_dependency 'uuidtools' end
334d3daf95eb8940163352ecb06ee5b3c372020b
thrifty-bunny.gemspec
thrifty-bunny.gemspec
Ruby
<|file_sep|>original/.idea/artifacts/ProgrammingLife_jar.xml <|file_sep|>current/.idea/artifacts/ProgrammingLife_jar.xml <|file_sep|>updated/.idea/artifacts/ProgrammingLife_jar.xml
<component name="ArtifactManager"> <artifact type="jar" name="ProgrammingLife:jar"> <output-path>$PROJECT_DIR$/out/artifacts/ProgrammingLife_jar</output-path> <root id="archive" name="ProgrammingLife.jar"> <element id="module-output" name="ProgrammingLife" /> <element id="extracted-dir" path="$MAVEN_REPOSITORY$/com/diffplug/durian/durian/3.4.0/durian-3.4.0.jar" path-in-jar="/" /> <element id="extracted-dir" path="$MAVEN_REPOSITORY$/org/apache/commons/commons-lang3/3.5/commons-lang3-3.5.jar" path-in-jar="/" /> </root> </artifact> </component>
<|file_sep|>original/.idea/artifacts/ProgrammingLife_jar.xml <|file_sep|>current/.idea/artifacts/ProgrammingLife_jar.xml <|file_sep|>updated/.idea/artifacts/ProgrammingLife_jar.xml <component name="ArtifactManager"> <artifact type="jar" name="ProgrammingLife:jar"> <output-path>$PROJECT_DIR$/out/artifacts/ProgrammingLife_jar</output-path> <root id="archive" name="ProgrammingLife.jar"> <element id="module-output" name="ProgrammingLife" /> <element id="extracted-dir" path="$MAVEN_REPOSITORY$/com/diffplug/durian/durian/3.4.0/durian-3.4.0.jar" path-in-jar="/" /> <element id="extracted-dir" path="$MAVEN_REPOSITORY$/org/apache/commons/commons-lang3/3.5/commons-lang3-3.5.jar" path-in-jar="/" /> </root> </artifact> </component>
8d12bc080ef9f976533434f78a674578bdeef189
.idea/artifacts/ProgrammingLife_jar.xml
.idea/artifacts/ProgrammingLife_jar.xml
XML
<|file_sep|>app/Console/Kernel.php.diff original: ->weeklyOn(5, '4:00'); updated: ->weeklyOn(6, '4:00'); // 6 = Saturday <|file_sep|>original/app/Console/Kernel.php /** * Define the application's command schedule. * * @param \Illuminate\Console\Scheduling\Schedule $schedule * @return void */ protected function schedule(Schedule $schedule) { // Synchronise student list with the LDAP $schedule->command('syncstudents') ->weeklyOn(5, '4:00'); // Send notification mails $schedule->command('sendnotificationmails --subscribed=instantly') ->everyThirtyMinutes(); $schedule->command('sendnotificationmails --subscribed=daily') ->dailyAt('4:00'); $schedule->command('sendnotificationmails --subscribed=weekly') ->weeklyOn(6, '4:00'); } } <|file_sep|>current/app/Console/Kernel.php /** * Define the application's command schedule. * * @param \Illuminate\Console\Scheduling\Schedule $schedule * @return void */ protected function schedule(Schedule $schedule) { // Synchronise student list with the LDAP $schedule->command('syncstudents') ->weeklyOn(6, '4:00'); // 6 = Saturday // Send notification mails $schedule->command('sendnotificationmails --subscribed=instantly') ->everyThirtyMinutes(); $schedule->command('sendnotificationmails --subscribed=daily') ->dailyAt('4:00'); $schedule->command('sendnotificationmails --subscribed=weekly') ->weeklyOn(6, '4:00'); } } <|file_sep|>updated/app/Console/Kernel.php
/** * Define the application's command schedule. * * @param \Illuminate\Console\Scheduling\Schedule $schedule * @return void */ protected function schedule(Schedule $schedule) { // Synchronise student list with the LDAP $schedule->command('syncstudents') ->weeklyOn(6, '4:00'); // 6 = Saturday // Send notification mails $schedule->command('sendnotificationmails --subscribed=instantly') ->everyThirtyMinutes(); $schedule->command('sendnotificationmails --subscribed=daily') ->dailyAt('4:00'); $schedule->command('sendnotificationmails --subscribed=weekly') ->weeklyOn(0, '4:00'); // 0 = Sunday } }
<|file_sep|>app/Console/Kernel.php.diff original: ->weeklyOn(5, '4:00'); updated: ->weeklyOn(6, '4:00'); // 6 = Saturday <|file_sep|>original/app/Console/Kernel.php /** * Define the application's command schedule. * * @param \Illuminate\Console\Scheduling\Schedule $schedule * @return void */ protected function schedule(Schedule $schedule) { // Synchronise student list with the LDAP $schedule->command('syncstudents') ->weeklyOn(5, '4:00'); // Send notification mails $schedule->command('sendnotificationmails --subscribed=instantly') ->everyThirtyMinutes(); $schedule->command('sendnotificationmails --subscribed=daily') ->dailyAt('4:00'); $schedule->command('sendnotificationmails --subscribed=weekly') ->weeklyOn(6, '4:00'); } } <|file_sep|>current/app/Console/Kernel.php /** * Define the application's command schedule. * * @param \Illuminate\Console\Scheduling\Schedule $schedule * @return void */ protected function schedule(Schedule $schedule) { // Synchronise student list with the LDAP $schedule->command('syncstudents') ->weeklyOn(6, '4:00'); // 6 = Saturday // Send notification mails $schedule->command('sendnotificationmails --subscribed=instantly') ->everyThirtyMinutes(); $schedule->command('sendnotificationmails --subscribed=daily') ->dailyAt('4:00'); $schedule->command('sendnotificationmails --subscribed=weekly') ->weeklyOn(6, '4:00'); } } <|file_sep|>updated/app/Console/Kernel.php /** * Define the application's command schedule. * * @param \Illuminate\Console\Scheduling\Schedule $schedule * @return void */ protected function schedule(Schedule $schedule) { // Synchronise student list with the LDAP $schedule->command('syncstudents') ->weeklyOn(6, '4:00'); // 6 = Saturday // Send notification mails $schedule->command('sendnotificationmails --subscribed=instantly') ->everyThirtyMinutes(); $schedule->command('sendnotificationmails --subscribed=daily') ->dailyAt('4:00'); $schedule->command('sendnotificationmails --subscribed=weekly') ->weeklyOn(0, '4:00'); // 0 = Sunday } }
21d3ccd3493a340fa0168f17df9ce2baab7f1347
app/Console/Kernel.php
app/Console/Kernel.php
PHP
<|file_sep|>original/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb describe Rubycritic::SmellAdapter::Flog do before do sample_path = "test/samples/flog/smelly.rb" flog = Rubycritic::Analyser::Flog.new([sample_path]) @adapter = Rubycritic::SmellAdapter::Flog.new(flog) end it "detects smells" do @adapter.smells.wont_be_empty end it "has smells with messages" do smell = @adapter.smells.first smell.message.must_be_kind_of String end it "has smells with scores" do smell = @adapter.smells.first smell.score.must_be_kind_of Numeric end end <|file_sep|>current/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb describe Rubycritic::SmellAdapter::Flog do before do sample_path = "test/samples/flog/smelly.rb" flog = Rubycritic::Analyser::Flog.new([sample_path]) @adapter = Rubycritic::SmellAdapter::Flog.new(flog) end it "detects smells" do @adapter.smells.wont_be_empty end it "has smells with messages" do smell = @adapter.smells.first smell.message.must_be_kind_of String end it "has smells with scores" do smell = @adapter.smells.first smell.score.must_be_kind_of Numeric end end <|file_sep|>updated/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
describe Rubycritic::SmellAdapter::Flog do before do sample_path = "test/samples/flog/smelly.rb" flog = Rubycritic::Analyser::Flog.new([sample_path]) @adapter = Rubycritic::SmellAdapter::Flog.new(flog) end it "detects smells" do @adapter.smells.wont_be_empty end it "has smells with messages" do smell = @adapter.smells.first smell.message.must_be_instance_of String end it "has smells with scores" do smell = @adapter.smells.first smell.score.must_be_kind_of Numeric end end
<|file_sep|>original/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb describe Rubycritic::SmellAdapter::Flog do before do sample_path = "test/samples/flog/smelly.rb" flog = Rubycritic::Analyser::Flog.new([sample_path]) @adapter = Rubycritic::SmellAdapter::Flog.new(flog) end it "detects smells" do @adapter.smells.wont_be_empty end it "has smells with messages" do smell = @adapter.smells.first smell.message.must_be_kind_of String end it "has smells with scores" do smell = @adapter.smells.first smell.score.must_be_kind_of Numeric end end <|file_sep|>current/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb describe Rubycritic::SmellAdapter::Flog do before do sample_path = "test/samples/flog/smelly.rb" flog = Rubycritic::Analyser::Flog.new([sample_path]) @adapter = Rubycritic::SmellAdapter::Flog.new(flog) end it "detects smells" do @adapter.smells.wont_be_empty end it "has smells with messages" do smell = @adapter.smells.first smell.message.must_be_kind_of String end it "has smells with scores" do smell = @adapter.smells.first smell.score.must_be_kind_of Numeric end end <|file_sep|>updated/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb describe Rubycritic::SmellAdapter::Flog do before do sample_path = "test/samples/flog/smelly.rb" flog = Rubycritic::Analyser::Flog.new([sample_path]) @adapter = Rubycritic::SmellAdapter::Flog.new(flog) end it "detects smells" do @adapter.smells.wont_be_empty end it "has smells with messages" do smell = @adapter.smells.first smell.message.must_be_instance_of String end it "has smells with scores" do smell = @adapter.smells.first smell.score.must_be_kind_of Numeric end end
e1da8c877977b5a6d5a42aa6d7d88694b2e92e34
test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
Ruby
<|file_sep|>original/hieradata/common.yaml --- message: "This node is using common data" #Puppet Server Tuning puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0 <|file_sep|>current/hieradata/common.yaml --- message: "This node is using common data" #Puppet Server Tuning puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0 <|file_sep|>updated/hieradata/common.yaml
--- message: "This node is using common data"
<|file_sep|>original/hieradata/common.yaml --- message: "This node is using common data" #Puppet Server Tuning puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0 <|file_sep|>current/hieradata/common.yaml --- message: "This node is using common data" #Puppet Server Tuning puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0 <|file_sep|>updated/hieradata/common.yaml --- message: "This node is using common data"
5497d199b3c851a317638905e77f44f2fafc64d4
hieradata/common.yaml
hieradata/common.yaml
YAML
<|file_sep|>original/packages/opencv/opencv-samples_git.bb <|file_sep|>current/packages/opencv/opencv-samples_git.bb <|file_sep|>updated/packages/opencv/opencv-samples_git.bb
DESCRIPTION = "Opencv : The Open Computer Vision Library" HOMEPAGE = "http://sourceforge.net/projects/opencvlibrary" SECTION = "libs" PRIORITY = "optional" LICENSE = "GPLv2" PR = "r3" DEPENDS = "opencv" SRCREV = "fb7cc87ed6a7425f3edc58e2d1eb44599496ac09" SRC_URI = "git://github.com/nzjrs/opencv.git;protocol=git;branch=libv4l" PV = "1.1.0+git${SRCREV}" S = "${WORKDIR}/git" do_install() { cd samples/c install -d ${D}/${bindir} for i in *.c; do
<|file_sep|>original/packages/opencv/opencv-samples_git.bb <|file_sep|>current/packages/opencv/opencv-samples_git.bb <|file_sep|>updated/packages/opencv/opencv-samples_git.bb DESCRIPTION = "Opencv : The Open Computer Vision Library" HOMEPAGE = "http://sourceforge.net/projects/opencvlibrary" SECTION = "libs" PRIORITY = "optional" LICENSE = "GPLv2" PR = "r3" DEPENDS = "opencv" SRCREV = "fb7cc87ed6a7425f3edc58e2d1eb44599496ac09" SRC_URI = "git://github.com/nzjrs/opencv.git;protocol=git;branch=libv4l" PV = "1.1.0+git${SRCREV}" S = "${WORKDIR}/git" do_install() { cd samples/c install -d ${D}/${bindir} for i in *.c; do
3c871af940959b5543908e909d7747ba0b4affaa
packages/opencv/opencv-samples_git.bb
packages/opencv/opencv-samples_git.bb
BitBake
<|file_sep|>original/LICENSE.txt Copyright (c) 2015 Bouke van der Bijl MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION <|file_sep|>current/LICENSE.txt Copyright (c) 2015 Bouke van der Bijl MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION <|file_sep|>updated/LICENSE.txt
Copyright (c) 2015 Shopify MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
<|file_sep|>original/LICENSE.txt Copyright (c) 2015 Bouke van der Bijl MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION <|file_sep|>current/LICENSE.txt Copyright (c) 2015 Bouke van der Bijl MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION <|file_sep|>updated/LICENSE.txt Copyright (c) 2015 Shopify MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
336086d5b6c8a5289aa9a3cdd3751118d1f087dd
LICENSE.txt
LICENSE.txt
Text
<|file_sep|>setup/optional_setup_steps.sh.diff original: updated: # Dependencies for Sublime Text (from https://www.sublimetext.com/docs/3/linux_repositories.html#apt) wget -qO - https://download.sublimetext.com/sublimehq-pub.gpg | sudo apt-key add - sudo apt-get install -yq apt-transport-https echo "deb https://download.sublimetext.com/ apt/dev/" | sudo tee /etc/apt/sources.list.d/sublime-text.list sudo apt-get update <|file_sep|>setup/optional_setup_steps.sh.diff original: updated: sublime-text <|file_sep|>original/setup/optional_setup_steps.sh terminator \ vim \ wget # Allow SSH into this machine sudo apt-get install -yq openssh-server # Set the default settings for gedit to be OK for programming gsettings set org.gnome.gedit.preferences.editor auto-indent true gsettings set org.gnome.gedit.preferences.editor bracket-matching true gsettings set org.gnome.gedit.preferences.editor display-line-numbers true gsettings set org.gnome.gedit.preferences.editor highlight-current-line true gsettings set org.gnome.gedit.preferences.editor insert-spaces true gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2" # Clean up the Unity launcher gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']" # Set the cache to timeout after 1 hour (setting is in seconds) git config --global credential.helper 'cache --timeout=3600' <|file_sep|>current/setup/optional_setup_steps.sh emacs \ gitk \ sublime-text terminator \ vim \ wget # Allow SSH into this machine sudo apt-get install -yq openssh-server # Set the default settings for gedit to be OK for programming gsettings set org.gnome.gedit.preferences.editor auto-indent true gsettings set org.gnome.gedit.preferences.editor bracket-matching true gsettings set org.gnome.gedit.preferences.editor display-line-numbers true gsettings set org.gnome.gedit.preferences.editor highlight-current-line true gsettings set org.gnome.gedit.preferences.editor insert-spaces true gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2" # Clean up the Unity launcher gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']" <|file_sep|>updated/setup/optional_setup_steps.sh
emacs \ gitk \ sublime-text terminator \ vim \ wget # Allow SSH into this machine sudo apt-get install -yq openssh-server # Set the default settings for gedit to be OK for programming gsettings set org.gnome.gedit.preferences.editor auto-indent true gsettings set org.gnome.gedit.preferences.editor bracket-matching true gsettings set org.gnome.gedit.preferences.editor display-line-numbers true gsettings set org.gnome.gedit.preferences.editor highlight-current-line true gsettings set org.gnome.gedit.preferences.editor insert-spaces true gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2" # Clean up the Unity launcher gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
<|file_sep|>setup/optional_setup_steps.sh.diff original: updated: # Dependencies for Sublime Text (from https://www.sublimetext.com/docs/3/linux_repositories.html#apt) wget -qO - https://download.sublimetext.com/sublimehq-pub.gpg | sudo apt-key add - sudo apt-get install -yq apt-transport-https echo "deb https://download.sublimetext.com/ apt/dev/" | sudo tee /etc/apt/sources.list.d/sublime-text.list sudo apt-get update <|file_sep|>setup/optional_setup_steps.sh.diff original: updated: sublime-text <|file_sep|>original/setup/optional_setup_steps.sh terminator \ vim \ wget # Allow SSH into this machine sudo apt-get install -yq openssh-server # Set the default settings for gedit to be OK for programming gsettings set org.gnome.gedit.preferences.editor auto-indent true gsettings set org.gnome.gedit.preferences.editor bracket-matching true gsettings set org.gnome.gedit.preferences.editor display-line-numbers true gsettings set org.gnome.gedit.preferences.editor highlight-current-line true gsettings set org.gnome.gedit.preferences.editor insert-spaces true gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2" # Clean up the Unity launcher gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']" # Set the cache to timeout after 1 hour (setting is in seconds) git config --global credential.helper 'cache --timeout=3600' <|file_sep|>current/setup/optional_setup_steps.sh emacs \ gitk \ sublime-text terminator \ vim \ wget # Allow SSH into this machine sudo apt-get install -yq openssh-server # Set the default settings for gedit to be OK for programming gsettings set org.gnome.gedit.preferences.editor auto-indent true gsettings set org.gnome.gedit.preferences.editor bracket-matching true gsettings set org.gnome.gedit.preferences.editor display-line-numbers true gsettings set org.gnome.gedit.preferences.editor highlight-current-line true gsettings set org.gnome.gedit.preferences.editor insert-spaces true gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2" # Clean up the Unity launcher gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']" <|file_sep|>updated/setup/optional_setup_steps.sh emacs \ gitk \ sublime-text terminator \ vim \ wget # Allow SSH into this machine sudo apt-get install -yq openssh-server # Set the default settings for gedit to be OK for programming gsettings set org.gnome.gedit.preferences.editor auto-indent true gsettings set org.gnome.gedit.preferences.editor bracket-matching true gsettings set org.gnome.gedit.preferences.editor display-line-numbers true gsettings set org.gnome.gedit.preferences.editor highlight-current-line true gsettings set org.gnome.gedit.preferences.editor insert-spaces true gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2" # Clean up the Unity launcher gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
e63b0192ed841977faec8dd321edfae27e7c8fa2
setup/optional_setup_steps.sh
setup/optional_setup_steps.sh
Shell
<|file_sep|>original/src/com/kduda/battleships/Main.java import javafx.application.Application; import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; public class Main extends Application { public static void main(String[] args) { launch(args); } @Override public void start(Stage primaryStage) throws Exception { Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml")); // setUserAgentStylesheet(STYLESHEET_CASPIAN); primaryStage.setTitle("Battleships"); primaryStage.setScene(new Scene(root,1024,768)); primaryStage.show(); } } <|file_sep|>current/src/com/kduda/battleships/Main.java import javafx.application.Application; import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; public class Main extends Application { public static void main(String[] args) { launch(args); } @Override public void start(Stage primaryStage) throws Exception { Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml")); // setUserAgentStylesheet(STYLESHEET_CASPIAN); primaryStage.setTitle("Battleships"); primaryStage.setScene(new Scene(root,1024,768)); primaryStage.show(); } } <|file_sep|>updated/src/com/kduda/battleships/Main.java
import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; public class Main extends Application { public static void main(String[] args) { launch(args); } @Override public void start(Stage primaryStage) throws Exception { Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml")); // setUserAgentStylesheet(STYLESHEET_CASPIAN); primaryStage.setTitle("Battleships"); primaryStage.setScene(new Scene(root); primaryStage.setResizable(false); primaryStage.show(); } }
<|file_sep|>original/src/com/kduda/battleships/Main.java import javafx.application.Application; import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; public class Main extends Application { public static void main(String[] args) { launch(args); } @Override public void start(Stage primaryStage) throws Exception { Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml")); // setUserAgentStylesheet(STYLESHEET_CASPIAN); primaryStage.setTitle("Battleships"); primaryStage.setScene(new Scene(root,1024,768)); primaryStage.show(); } } <|file_sep|>current/src/com/kduda/battleships/Main.java import javafx.application.Application; import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; public class Main extends Application { public static void main(String[] args) { launch(args); } @Override public void start(Stage primaryStage) throws Exception { Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml")); // setUserAgentStylesheet(STYLESHEET_CASPIAN); primaryStage.setTitle("Battleships"); primaryStage.setScene(new Scene(root,1024,768)); primaryStage.show(); } } <|file_sep|>updated/src/com/kduda/battleships/Main.java import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; public class Main extends Application { public static void main(String[] args) { launch(args); } @Override public void start(Stage primaryStage) throws Exception { Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml")); // setUserAgentStylesheet(STYLESHEET_CASPIAN); primaryStage.setTitle("Battleships"); primaryStage.setScene(new Scene(root); primaryStage.setResizable(false); primaryStage.show(); } }
4a351c628fe8e9c16132c02592936232c54093a1
src/com/kduda/battleships/Main.java
src/com/kduda/battleships/Main.java
Java
<|file_sep|>original/Formula/php70-uuid.rb <|file_sep|>current/Formula/php70-uuid.rb <|file_sep|>updated/Formula/php70-uuid.rb
require File.expand_path("../../Abstract/abstract-php-extension", __FILE__) class Php70Uuid < AbstractPhp70Extension init desc "UUID extension" homepage "https://pecl.php.net/package/uuid" stable do url "https://pecl.php.net/get/uuid-1.0.4.tgz" sha256 "63079b6a62a9d43691ecbcd4eb52e5e5fe17b5a3d0f8e46e3c17ff265c06a11f" patch do # let's fix the path to uuid.h (uuid/uuid.h on linux, ossp/uuid.h on OSX) # uuid_mac & uuid_time might not be available on OSX, let's add test to avoid compiling issue on these functions url "https://gist.githubusercontent.com/romainneutron/fe068c297413aee565d5/raw/28d6ba0b6e902e82e71bb9a1ed768c836a8161e4/php-uuid-1.0.4" sha256 "5f0664d5c4f55d4f6c037dab9f198e697afa3f9266854ed3945d7697fdb692b2" end end head do url "https://git.php.net/repository/pecl/networking/uuid.git"
<|file_sep|>original/Formula/php70-uuid.rb <|file_sep|>current/Formula/php70-uuid.rb <|file_sep|>updated/Formula/php70-uuid.rb require File.expand_path("../../Abstract/abstract-php-extension", __FILE__) class Php70Uuid < AbstractPhp70Extension init desc "UUID extension" homepage "https://pecl.php.net/package/uuid" stable do url "https://pecl.php.net/get/uuid-1.0.4.tgz" sha256 "63079b6a62a9d43691ecbcd4eb52e5e5fe17b5a3d0f8e46e3c17ff265c06a11f" patch do # let's fix the path to uuid.h (uuid/uuid.h on linux, ossp/uuid.h on OSX) # uuid_mac & uuid_time might not be available on OSX, let's add test to avoid compiling issue on these functions url "https://gist.githubusercontent.com/romainneutron/fe068c297413aee565d5/raw/28d6ba0b6e902e82e71bb9a1ed768c836a8161e4/php-uuid-1.0.4" sha256 "5f0664d5c4f55d4f6c037dab9f198e697afa3f9266854ed3945d7697fdb692b2" end end head do url "https://git.php.net/repository/pecl/networking/uuid.git"
25c265f64b7c26bfd68c12c8179fa631b89c3617
Formula/php70-uuid.rb
Formula/php70-uuid.rb
Ruby
<|file_sep|>original/rules/management/commands/CorpusLogicUpdate.py <|file_sep|>current/rules/management/commands/CorpusLogicUpdate.py <|file_sep|>updated/rules/management/commands/CorpusLogicUpdate.py
import logging from django.core.management.base import BaseCommand, CommandError from plyara import YaraParser from rules.models import YaraRule # Configure Logging logging.basicConfig(level=logging.INFO) class Command(BaseCommand): help = 'Recalculate the logic hashes of the entire rule corpus' def handle(self, *args, **options): corpus = YaraRule.objects.all() rule_count = corpus.count() message = 'Updating logic hashes for {} rules'.format(rule_count) logging.info(message) rule_index = 0
<|file_sep|>original/rules/management/commands/CorpusLogicUpdate.py <|file_sep|>current/rules/management/commands/CorpusLogicUpdate.py <|file_sep|>updated/rules/management/commands/CorpusLogicUpdate.py import logging from django.core.management.base import BaseCommand, CommandError from plyara import YaraParser from rules.models import YaraRule # Configure Logging logging.basicConfig(level=logging.INFO) class Command(BaseCommand): help = 'Recalculate the logic hashes of the entire rule corpus' def handle(self, *args, **options): corpus = YaraRule.objects.all() rule_count = corpus.count() message = 'Updating logic hashes for {} rules'.format(rule_count) logging.info(message) rule_index = 0
b3f8be5b6ab7e4e713004447a3cfbda743d80394
rules/management/commands/CorpusLogicUpdate.py
rules/management/commands/CorpusLogicUpdate.py
Python
<|file_sep|>original/scripts/roles/cikit-php/tasks/xdebug.yml --- - name: Set name of directory for XDebug set_fact: cikit_php_xdebug: "xdebug-{{ php.xdebug }}" - name: Download XDebug get_url: url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz" dest: "/tmp/{{ cikit_php_xdebug }}.tgz" - name: Unarchive XDebug shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/" args: # Unarchive module does not work properly in Ansible 2.1.1.0. warn: no - include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}" <|file_sep|>current/scripts/roles/cikit-php/tasks/xdebug.yml --- - name: Set name of directory for XDebug set_fact: cikit_php_xdebug: "xdebug-{{ php.xdebug }}" - name: Download XDebug get_url: url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz" dest: "/tmp/{{ cikit_php_xdebug }}.tgz" - name: Unarchive XDebug shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/" args: # Unarchive module does not work properly in Ansible 2.1.1.0. warn: no - include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}" <|file_sep|>updated/scripts/roles/cikit-php/tasks/xdebug.yml
get_url: url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz" dest: "/tmp/{{ cikit_php_xdebug }}.tgz" - name: Unarchive XDebug shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/" args: # Unarchive module does not work properly in Ansible 2.1.1.0. warn: no - include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}" - name: Set environment variable lineinfile: dest: /etc/bash.bashrc line: 'export {{ item.key }}="{{ item.value }}"' regexp: "{{ item.value | regex_escape }}" insertbefore: BOF with_dict: XDEBUG_CONFIG: "idekey={{ php.extensions.xdebug.config.idekey }}" PHP_IDE_CONFIG: "serverName={{ ansible_fqdn }}"
<|file_sep|>original/scripts/roles/cikit-php/tasks/xdebug.yml --- - name: Set name of directory for XDebug set_fact: cikit_php_xdebug: "xdebug-{{ php.xdebug }}" - name: Download XDebug get_url: url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz" dest: "/tmp/{{ cikit_php_xdebug }}.tgz" - name: Unarchive XDebug shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/" args: # Unarchive module does not work properly in Ansible 2.1.1.0. warn: no - include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}" <|file_sep|>current/scripts/roles/cikit-php/tasks/xdebug.yml --- - name: Set name of directory for XDebug set_fact: cikit_php_xdebug: "xdebug-{{ php.xdebug }}" - name: Download XDebug get_url: url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz" dest: "/tmp/{{ cikit_php_xdebug }}.tgz" - name: Unarchive XDebug shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/" args: # Unarchive module does not work properly in Ansible 2.1.1.0. warn: no - include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}" <|file_sep|>updated/scripts/roles/cikit-php/tasks/xdebug.yml get_url: url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz" dest: "/tmp/{{ cikit_php_xdebug }}.tgz" - name: Unarchive XDebug shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/" args: # Unarchive module does not work properly in Ansible 2.1.1.0. warn: no - include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}" - name: Set environment variable lineinfile: dest: /etc/bash.bashrc line: 'export {{ item.key }}="{{ item.value }}"' regexp: "{{ item.value | regex_escape }}" insertbefore: BOF with_dict: XDEBUG_CONFIG: "idekey={{ php.extensions.xdebug.config.idekey }}" PHP_IDE_CONFIG: "serverName={{ ansible_fqdn }}"
ad7610ebdb647411307b2353fb01f7bc9e97b8ad
scripts/roles/cikit-php/tasks/xdebug.yml
scripts/roles/cikit-php/tasks/xdebug.yml
YAML
<|file_sep|>recipes/_nginx.rb.diff original: service 'nginx' do action [:enable, :start] supports :restart => true, :start => true, :stop => true, :reload => true end updated: <|file_sep|>original/recipes/_nginx.rb include_recipe 'oc-graphite::_uwsgi' service 'nginx' do action [:enable, :start] supports :restart => true, :start => true, :stop => true, :reload => true end template '/etc/nginx/sites-available/graphite' do source 'nginx-graphite.erb' owner 'root' group 'root' mode 0644 notifies :reload, 'service[nginx]', :delayed end link '/etc/nginx/sites-enabled/graphite' do to '/etc/nginx/sites-available/graphite' end file '/etc/nginx/sites-enabled/default' do <|file_sep|>current/recipes/_nginx.rb include_recipe 'oc-graphite::_uwsgi' template '/etc/nginx/sites-available/graphite' do source 'nginx-graphite.erb' owner 'root' group 'root' mode 0644 notifies :reload, 'service[nginx]', :delayed end link '/etc/nginx/sites-enabled/graphite' do to '/etc/nginx/sites-available/graphite' end file '/etc/nginx/sites-enabled/default' do action :delete notifies :reload, 'service[nginx]', :delayed only_if { node['oc-graphite']['nginx']['disable_default_vhost'] } end <|file_sep|>updated/recipes/_nginx.rb
group 'root' mode 0644 notifies :reload, 'service[nginx]', :delayed end link '/etc/nginx/sites-enabled/graphite' do to '/etc/nginx/sites-available/graphite' end file '/etc/nginx/sites-enabled/default' do action :delete notifies :reload, 'service[nginx]', :delayed only_if { node['oc-graphite']['nginx']['disable_default_vhost'] } end service 'nginx' do action [:enable, :start] supports :restart => true, :reload => true end
<|file_sep|>recipes/_nginx.rb.diff original: service 'nginx' do action [:enable, :start] supports :restart => true, :start => true, :stop => true, :reload => true end updated: <|file_sep|>original/recipes/_nginx.rb include_recipe 'oc-graphite::_uwsgi' service 'nginx' do action [:enable, :start] supports :restart => true, :start => true, :stop => true, :reload => true end template '/etc/nginx/sites-available/graphite' do source 'nginx-graphite.erb' owner 'root' group 'root' mode 0644 notifies :reload, 'service[nginx]', :delayed end link '/etc/nginx/sites-enabled/graphite' do to '/etc/nginx/sites-available/graphite' end file '/etc/nginx/sites-enabled/default' do <|file_sep|>current/recipes/_nginx.rb include_recipe 'oc-graphite::_uwsgi' template '/etc/nginx/sites-available/graphite' do source 'nginx-graphite.erb' owner 'root' group 'root' mode 0644 notifies :reload, 'service[nginx]', :delayed end link '/etc/nginx/sites-enabled/graphite' do to '/etc/nginx/sites-available/graphite' end file '/etc/nginx/sites-enabled/default' do action :delete notifies :reload, 'service[nginx]', :delayed only_if { node['oc-graphite']['nginx']['disable_default_vhost'] } end <|file_sep|>updated/recipes/_nginx.rb group 'root' mode 0644 notifies :reload, 'service[nginx]', :delayed end link '/etc/nginx/sites-enabled/graphite' do to '/etc/nginx/sites-available/graphite' end file '/etc/nginx/sites-enabled/default' do action :delete notifies :reload, 'service[nginx]', :delayed only_if { node['oc-graphite']['nginx']['disable_default_vhost'] } end service 'nginx' do action [:enable, :start] supports :restart => true, :reload => true end
765a9399a7716e7b85e3c6f9a343bb602044f9b3
recipes/_nginx.rb
recipes/_nginx.rb
Ruby
<|file_sep|>original/install-manpages.sh <|file_sep|>current/install-manpages.sh <|file_sep|>updated/install-manpages.sh
#!/bin/sh -e [ -z "$PREFIX" ] && PREFIX="/usr/local" [ -z "$MANDIR" ] && MANDIR="$PREFIX/share/man" install -d "$DESTDIR/$MANDIR/man1" for manpage in man/*; do install "$manpage" "$DESTDIR/$MANDIR/man1/`basename $manpage`" done
<|file_sep|>original/install-manpages.sh <|file_sep|>current/install-manpages.sh <|file_sep|>updated/install-manpages.sh #!/bin/sh -e [ -z "$PREFIX" ] && PREFIX="/usr/local" [ -z "$MANDIR" ] && MANDIR="$PREFIX/share/man" install -d "$DESTDIR/$MANDIR/man1" for manpage in man/*; do install "$manpage" "$DESTDIR/$MANDIR/man1/`basename $manpage`" done
2dc7414413bd2a17e7bb65a0cb1ba1ae8bf91dc9
install-manpages.sh
install-manpages.sh
Shell
<|file_sep|>original/tests/test_server.py from unittest.mock import ANY from queue_functions import do_work from server import handle_post from uploaders.s3 import get_url from uploaders.s3 import upload def test_post(): q = Mock() filename = 'afakefilename' files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]} hash_object = hashlib.md5(filename.encode()) audio_filename = hash_object.hexdigest() + "-" + filename analysis_filename = audio_filename + '.analysis.json' expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)} actual = json.reads(handle_post(q, files, get_url, upload)) q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload)) assert expected == actual <|file_sep|>current/tests/test_server.py from unittest.mock import ANY from queue_functions import do_work from server import handle_post from uploaders.s3 import get_url from uploaders.s3 import upload def test_post(): q = Mock() filename = 'afakefilename' files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]} hash_object = hashlib.md5(filename.encode()) audio_filename = hash_object.hexdigest() + "-" + filename analysis_filename = audio_filename + '.analysis.json' expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)} actual = json.reads(handle_post(q, files, get_url, upload)) q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload)) assert expected == actual <|file_sep|>updated/tests/test_server.py
from unittest.mock import ANY from queue_functions import do_work from server import handle_post from uploaders.s3 import get_url from uploaders.s3 import upload def test_post(): q = Mock() filename = 'afakefilename' files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]} hash_object = hashlib.md5(filename.encode()) audio_filename = hash_object.hexdigest() + "-" + filename analysis_filename = audio_filename + '.analysis.json' expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)} actual = json.loads(handle_post(q, files, get_url, upload)) q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload)) assert expected == actual
<|file_sep|>original/tests/test_server.py from unittest.mock import ANY from queue_functions import do_work from server import handle_post from uploaders.s3 import get_url from uploaders.s3 import upload def test_post(): q = Mock() filename = 'afakefilename' files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]} hash_object = hashlib.md5(filename.encode()) audio_filename = hash_object.hexdigest() + "-" + filename analysis_filename = audio_filename + '.analysis.json' expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)} actual = json.reads(handle_post(q, files, get_url, upload)) q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload)) assert expected == actual <|file_sep|>current/tests/test_server.py from unittest.mock import ANY from queue_functions import do_work from server import handle_post from uploaders.s3 import get_url from uploaders.s3 import upload def test_post(): q = Mock() filename = 'afakefilename' files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]} hash_object = hashlib.md5(filename.encode()) audio_filename = hash_object.hexdigest() + "-" + filename analysis_filename = audio_filename + '.analysis.json' expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)} actual = json.reads(handle_post(q, files, get_url, upload)) q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload)) assert expected == actual <|file_sep|>updated/tests/test_server.py from unittest.mock import ANY from queue_functions import do_work from server import handle_post from uploaders.s3 import get_url from uploaders.s3 import upload def test_post(): q = Mock() filename = 'afakefilename' files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]} hash_object = hashlib.md5(filename.encode()) audio_filename = hash_object.hexdigest() + "-" + filename analysis_filename = audio_filename + '.analysis.json' expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)} actual = json.loads(handle_post(q, files, get_url, upload)) q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload)) assert expected == actual
3b7dcc4d2a19b5ac03eebae35600c25dd038fe33
tests/test_server.py
tests/test_server.py
Python
<|file_sep|>original/perftest.py <|file_sep|>current/perftest.py <|file_sep|>updated/perftest.py
""" Simple peformance tests. """ import sys import time import couchdb def main(): print 'sys.version : %r' % (sys.version,) print 'sys.platform : %r' % (sys.platform,) tests = [create_doc, create_bulk_docs] if len(sys.argv) > 1: tests = [test for test in tests if test.__name__ in sys.argv[1:]] server = couchdb.Server() for test in tests:
<|file_sep|>original/perftest.py <|file_sep|>current/perftest.py <|file_sep|>updated/perftest.py """ Simple peformance tests. """ import sys import time import couchdb def main(): print 'sys.version : %r' % (sys.version,) print 'sys.platform : %r' % (sys.platform,) tests = [create_doc, create_bulk_docs] if len(sys.argv) > 1: tests = [test for test in tests if test.__name__ in sys.argv[1:]] server = couchdb.Server() for test in tests:
877406927bc4754daeab10b9bfb0f7879e8f6092
perftest.py
perftest.py
Python
<|file_sep|>original/tests/testMustache.php <|file_sep|>current/tests/testMustache.php <|file_sep|>updated/tests/testMustache.php
<?php /* * Mustache PHP Compiler - Test the Mustache class * * (c) July 2011 - Manuel Odendahl - wesen@ruinwesen.com */ require_once(dirname(__FILE__)."/../vendor/simpletest/autorun.php"); require_once(dirname(__FILE__)."/../Mustache.php"); function __filename($file) { return dirname(__FILE__).'/files/'.$file; } class TestMustache extends UnitTestCase { function testPartial() { $m = new Mustache(); $res = $m->partial(__filename("testPartial.mustache")); $this->assertEqual($res, "partial {{mustache}}\n");
<|file_sep|>original/tests/testMustache.php <|file_sep|>current/tests/testMustache.php <|file_sep|>updated/tests/testMustache.php <?php /* * Mustache PHP Compiler - Test the Mustache class * * (c) July 2011 - Manuel Odendahl - wesen@ruinwesen.com */ require_once(dirname(__FILE__)."/../vendor/simpletest/autorun.php"); require_once(dirname(__FILE__)."/../Mustache.php"); function __filename($file) { return dirname(__FILE__).'/files/'.$file; } class TestMustache extends UnitTestCase { function testPartial() { $m = new Mustache(); $res = $m->partial(__filename("testPartial.mustache")); $this->assertEqual($res, "partial {{mustache}}\n");
266132d339ca83c84b54841ff14804a6932289ae
tests/testMustache.php
tests/testMustache.php
PHP
<|file_sep|>original/.travis.yml language: go go: - 1.6.x - 1.7.x - 1.8.x - 1.9.x - 1.10.x - 1.11.x - 1.12.x - tip gobuild_args: -tags no_pkg_config dist: xenial addons: apt: packages: - bison - flex - automake - autoconf - libtool - make <|file_sep|>current/.travis.yml language: go go: - 1.6.x - 1.7.x - 1.8.x - 1.9.x - 1.10.x - 1.11.x - 1.12.x - tip gobuild_args: -tags no_pkg_config dist: xenial addons: apt: packages: - bison - flex - automake - autoconf - libtool - make <|file_sep|>updated/.travis.yml
language: go go: - 1.7.x - 1.8.x - 1.9.x - 1.10.x - 1.11.x - 1.12.x - tip gobuild_args: -tags no_pkg_config dist: xenial addons: apt: packages: - bison - flex - automake - autoconf - libtool - make - gcc
<|file_sep|>original/.travis.yml language: go go: - 1.6.x - 1.7.x - 1.8.x - 1.9.x - 1.10.x - 1.11.x - 1.12.x - tip gobuild_args: -tags no_pkg_config dist: xenial addons: apt: packages: - bison - flex - automake - autoconf - libtool - make <|file_sep|>current/.travis.yml language: go go: - 1.6.x - 1.7.x - 1.8.x - 1.9.x - 1.10.x - 1.11.x - 1.12.x - tip gobuild_args: -tags no_pkg_config dist: xenial addons: apt: packages: - bison - flex - automake - autoconf - libtool - make <|file_sep|>updated/.travis.yml language: go go: - 1.7.x - 1.8.x - 1.9.x - 1.10.x - 1.11.x - 1.12.x - tip gobuild_args: -tags no_pkg_config dist: xenial addons: apt: packages: - bison - flex - automake - autoconf - libtool - make - gcc
ce52fb9bf92529e068f200314fd6cbbbb251c7c2
.travis.yml
.travis.yml
YAML
<|file_sep|>original/templates/ubuntu-14.04-server-amd64/vbox.sh # Without libdbus virtualbox would not start automatically after compile apt-get -y install --no-install-recommends libdbus-1-3 # Remove existing VirtualBox guest additions /etc/init.d/virtualbox-ose-guest-utils stop rmmod vboxguest aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils aptitude -y install dkms # Install the VirtualBox guest additions VBOX_VERSION=$(cat /home/vagrant/.vbox_version) VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso mount -o loop $VBOX_ISO /mnt yes|sh /mnt/VBoxLinuxAdditions.run umount /mnt # Cleanup rm $VBOX_ISO <|file_sep|>current/templates/ubuntu-14.04-server-amd64/vbox.sh # Without libdbus virtualbox would not start automatically after compile apt-get -y install --no-install-recommends libdbus-1-3 # Remove existing VirtualBox guest additions /etc/init.d/virtualbox-ose-guest-utils stop rmmod vboxguest aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils aptitude -y install dkms # Install the VirtualBox guest additions VBOX_VERSION=$(cat /home/vagrant/.vbox_version) VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso mount -o loop $VBOX_ISO /mnt yes|sh /mnt/VBoxLinuxAdditions.run umount /mnt # Cleanup rm $VBOX_ISO <|file_sep|>updated/templates/ubuntu-14.04-server-amd64/vbox.sh
apt-get -y install --no-install-recommends libdbus-1-3 # Remove existing VirtualBox guest additions /etc/init.d/virtualbox-ose-guest-utils stop rmmod vboxguest aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils aptitude -y install dkms # Install the VirtualBox guest additions VBOX_VERSION=$(cat /home/vagrant/.vbox_version) VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso mount -o loop $VBOX_ISO /mnt yes|sh /mnt/VBoxLinuxAdditions.run umount /mnt # Temporary fix for VirtualBox Additions version 4.3.10 # issue #12879, see https://www.virtualbox.org/ticket/12879 [ -e /usr/lib/VBoxGuestAdditions ] || ln -s /opt/VBoxGuestAdditions-$VBOX_VERSION/lib/VBoxGuestAdditions /usr/lib/VBoxGuestAdditions # Cleanup rm $VBOX_ISO
<|file_sep|>original/templates/ubuntu-14.04-server-amd64/vbox.sh # Without libdbus virtualbox would not start automatically after compile apt-get -y install --no-install-recommends libdbus-1-3 # Remove existing VirtualBox guest additions /etc/init.d/virtualbox-ose-guest-utils stop rmmod vboxguest aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils aptitude -y install dkms # Install the VirtualBox guest additions VBOX_VERSION=$(cat /home/vagrant/.vbox_version) VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso mount -o loop $VBOX_ISO /mnt yes|sh /mnt/VBoxLinuxAdditions.run umount /mnt # Cleanup rm $VBOX_ISO <|file_sep|>current/templates/ubuntu-14.04-server-amd64/vbox.sh # Without libdbus virtualbox would not start automatically after compile apt-get -y install --no-install-recommends libdbus-1-3 # Remove existing VirtualBox guest additions /etc/init.d/virtualbox-ose-guest-utils stop rmmod vboxguest aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils aptitude -y install dkms # Install the VirtualBox guest additions VBOX_VERSION=$(cat /home/vagrant/.vbox_version) VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso mount -o loop $VBOX_ISO /mnt yes|sh /mnt/VBoxLinuxAdditions.run umount /mnt # Cleanup rm $VBOX_ISO <|file_sep|>updated/templates/ubuntu-14.04-server-amd64/vbox.sh apt-get -y install --no-install-recommends libdbus-1-3 # Remove existing VirtualBox guest additions /etc/init.d/virtualbox-ose-guest-utils stop rmmod vboxguest aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils aptitude -y install dkms # Install the VirtualBox guest additions VBOX_VERSION=$(cat /home/vagrant/.vbox_version) VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso mount -o loop $VBOX_ISO /mnt yes|sh /mnt/VBoxLinuxAdditions.run umount /mnt # Temporary fix for VirtualBox Additions version 4.3.10 # issue #12879, see https://www.virtualbox.org/ticket/12879 [ -e /usr/lib/VBoxGuestAdditions ] || ln -s /opt/VBoxGuestAdditions-$VBOX_VERSION/lib/VBoxGuestAdditions /usr/lib/VBoxGuestAdditions # Cleanup rm $VBOX_ISO
3c0deaa545084ad95c0ce614b6ab01bfda2dcb0d
templates/ubuntu-14.04-server-amd64/vbox.sh
templates/ubuntu-14.04-server-amd64/vbox.sh
Shell
<|file_sep|>original/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php <?php namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge; interface MailSenderInterface { /** * @param MailMessage $message * @return void * @throws \Zend_Mail_Exception */ public function send(MailMessage $message); } <|file_sep|>current/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php <?php namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge; interface MailSenderInterface { /** * @param MailMessage $message * @return void * @throws \Zend_Mail_Exception */ public function send(MailMessage $message); } <|file_sep|>updated/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
<?php namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge; interface MailSenderInterface { /** * @param MailMessage $message * @return void */ public function send(MailMessage $message); }
<|file_sep|>original/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php <?php namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge; interface MailSenderInterface { /** * @param MailMessage $message * @return void * @throws \Zend_Mail_Exception */ public function send(MailMessage $message); } <|file_sep|>current/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php <?php namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge; interface MailSenderInterface { /** * @param MailMessage $message * @return void * @throws \Zend_Mail_Exception */ public function send(MailMessage $message); } <|file_sep|>updated/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php <?php namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge; interface MailSenderInterface { /** * @param MailMessage $message * @return void */ public function send(MailMessage $message); }
a2532cd8abb5adccf29b5136c4ba23ecf46d3b36
src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
PHP
<|file_sep|>original/package.json "type": "git", "url": "git://github.com/glaubinix/changelog-builder.git" }, "license": "MIT", "author": { "name": "Stephan Vock", "email": "stephan.vock@gmail.com" }, "keywords": [ "git", "svn", "changelog" ], "dependencies": { "jira": "git://github.com/Woellchen/node-jira#develop" }, "devDependencies": { "precommit-hook": "*", "tape": "^2.14.0" } } <|file_sep|>current/package.json "type": "git", "url": "git://github.com/glaubinix/changelog-builder.git" }, "license": "MIT", "author": { "name": "Stephan Vock", "email": "stephan.vock@gmail.com" }, "keywords": [ "git", "svn", "changelog" ], "dependencies": { "jira": "git://github.com/Woellchen/node-jira#develop" }, "devDependencies": { "precommit-hook": "*", "tape": "^2.14.0" } } <|file_sep|>updated/package.json
"type": "git", "url": "git://github.com/glaubinix/changelog-builder.git" }, "license": "MIT", "author": { "name": "Stephan Vock", "email": "stephan.vock@gmail.com" }, "keywords": [ "git", "svn", "changelog" ], "dependencies": { "jira": "^0.10.0" }, "devDependencies": { "precommit-hook": "*", "tape": "^2.14.0" } }
<|file_sep|>original/package.json "type": "git", "url": "git://github.com/glaubinix/changelog-builder.git" }, "license": "MIT", "author": { "name": "Stephan Vock", "email": "stephan.vock@gmail.com" }, "keywords": [ "git", "svn", "changelog" ], "dependencies": { "jira": "git://github.com/Woellchen/node-jira#develop" }, "devDependencies": { "precommit-hook": "*", "tape": "^2.14.0" } } <|file_sep|>current/package.json "type": "git", "url": "git://github.com/glaubinix/changelog-builder.git" }, "license": "MIT", "author": { "name": "Stephan Vock", "email": "stephan.vock@gmail.com" }, "keywords": [ "git", "svn", "changelog" ], "dependencies": { "jira": "git://github.com/Woellchen/node-jira#develop" }, "devDependencies": { "precommit-hook": "*", "tape": "^2.14.0" } } <|file_sep|>updated/package.json "type": "git", "url": "git://github.com/glaubinix/changelog-builder.git" }, "license": "MIT", "author": { "name": "Stephan Vock", "email": "stephan.vock@gmail.com" }, "keywords": [ "git", "svn", "changelog" ], "dependencies": { "jira": "^0.10.0" }, "devDependencies": { "precommit-hook": "*", "tape": "^2.14.0" } }
b6462822ad53c586fd737e016fc92b9487a1182b
package.json
package.json
JSON
<|file_sep|>original/composer.json "homepage": "http://mickael-vieira.com" } ], "require": { "php": ">=7.0", "psr/link": "^1.0" }, "require-dev": { "phpspec/phpspec": "^3.0", "squizlabs/php_codesniffer": "^3.0", "henrikbjorn/phpspec-code-coverage": "^3.0", "satooshi/php-coveralls": "^1.0" }, "config": { "bin-dir": "bin" }, "autoload": { "psr-4": { "CollectionJson\\": "src/CollectionJson/" } }, <|file_sep|>current/composer.json "homepage": "http://mickael-vieira.com" } ], "require": { "php": ">=7.0", "psr/link": "^1.0" }, "require-dev": { "phpspec/phpspec": "^3.0", "squizlabs/php_codesniffer": "^3.0", "henrikbjorn/phpspec-code-coverage": "^3.0", "satooshi/php-coveralls": "^1.0" }, "config": { "bin-dir": "bin" }, "autoload": { "psr-4": { "CollectionJson\\": "src/CollectionJson/" } }, <|file_sep|>updated/composer.json
"homepage": "http://mickael-vieira.com" } ], "require": { "php": ">=7.0", "psr/link": "^1.0" }, "require-dev": { "phpspec/phpspec": "^3.0", "squizlabs/php_codesniffer": "^3.0", "satooshi/php-coveralls": "^1.0", "leanphp/phpspec-code-coverage": "^3.1" }, "config": { "bin-dir": "bin" }, "autoload": { "psr-4": { "CollectionJson\\": "src/CollectionJson/" } },
<|file_sep|>original/composer.json "homepage": "http://mickael-vieira.com" } ], "require": { "php": ">=7.0", "psr/link": "^1.0" }, "require-dev": { "phpspec/phpspec": "^3.0", "squizlabs/php_codesniffer": "^3.0", "henrikbjorn/phpspec-code-coverage": "^3.0", "satooshi/php-coveralls": "^1.0" }, "config": { "bin-dir": "bin" }, "autoload": { "psr-4": { "CollectionJson\\": "src/CollectionJson/" } }, <|file_sep|>current/composer.json "homepage": "http://mickael-vieira.com" } ], "require": { "php": ">=7.0", "psr/link": "^1.0" }, "require-dev": { "phpspec/phpspec": "^3.0", "squizlabs/php_codesniffer": "^3.0", "henrikbjorn/phpspec-code-coverage": "^3.0", "satooshi/php-coveralls": "^1.0" }, "config": { "bin-dir": "bin" }, "autoload": { "psr-4": { "CollectionJson\\": "src/CollectionJson/" } }, <|file_sep|>updated/composer.json "homepage": "http://mickael-vieira.com" } ], "require": { "php": ">=7.0", "psr/link": "^1.0" }, "require-dev": { "phpspec/phpspec": "^3.0", "squizlabs/php_codesniffer": "^3.0", "satooshi/php-coveralls": "^1.0", "leanphp/phpspec-code-coverage": "^3.1" }, "config": { "bin-dir": "bin" }, "autoload": { "psr-4": { "CollectionJson\\": "src/CollectionJson/" } },
05b3d4913f7f95bf2d0a970a8a99d2656b3e2ea7
composer.json
composer.json
JSON
<|file_sep|>lib/learn_web/client/user/me.rb.diff original: exit updated: exit 1 <|file_sep|>lib/learn_web/client/user/me.rb.diff original: exit updated: exit 1 <|file_sep|>original/lib/learn_web/client/user/me.rb populate_attributes! elsif silent_output == false case response.status when 401 puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset" exit when 500 puts "Something went wrong. Please try again." exit else puts "Something went wrong. Please try again." exit end end self end end end end end <|file_sep|>current/lib/learn_web/client/user/me.rb populate_attributes! elsif silent_output == false case response.status when 401 puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset" exit 1 when 500 puts "Something went wrong. Please try again." exit 1 else puts "Something went wrong. Please try again." exit end end self end end end end end <|file_sep|>updated/lib/learn_web/client/user/me.rb
populate_attributes! elsif silent_output == false case response.status when 401 puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset" exit 1 when 500 puts "Something went wrong. Please try again." exit 1 else puts "Something went wrong. Please try again." exit 1 end end self end end end end end
<|file_sep|>lib/learn_web/client/user/me.rb.diff original: exit updated: exit 1 <|file_sep|>lib/learn_web/client/user/me.rb.diff original: exit updated: exit 1 <|file_sep|>original/lib/learn_web/client/user/me.rb populate_attributes! elsif silent_output == false case response.status when 401 puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset" exit when 500 puts "Something went wrong. Please try again." exit else puts "Something went wrong. Please try again." exit end end self end end end end end <|file_sep|>current/lib/learn_web/client/user/me.rb populate_attributes! elsif silent_output == false case response.status when 401 puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset" exit 1 when 500 puts "Something went wrong. Please try again." exit 1 else puts "Something went wrong. Please try again." exit end end self end end end end end <|file_sep|>updated/lib/learn_web/client/user/me.rb populate_attributes! elsif silent_output == false case response.status when 401 puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset" exit 1 when 500 puts "Something went wrong. Please try again." exit 1 else puts "Something went wrong. Please try again." exit 1 end end self end end end end end
d0ad83967a024eb655b048d2a5cd48db1d54469a
lib/learn_web/client/user/me.rb
lib/learn_web/client/user/me.rb
Ruby
<|file_sep|>data/building-hours/3-2-print-center.yaml.diff original: closedForChapelTime: true updated: # closedForChapelTime: true <|file_sep|>original/data/building-hours/3-2-print-center.yaml name: Print Center image: print category: Mail and Packages schedule: - title: Hours closedForChapelTime: true hours: # - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'} breakSchedule: fall: [] thanksgiving: [] winter: [] interim: [] spring: [] easter: [] summer: - title: Hours hours: - {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'} <|file_sep|>current/data/building-hours/3-2-print-center.yaml name: Print Center image: print category: Mail and Packages schedule: - title: Hours # closedForChapelTime: true hours: # - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'} breakSchedule: fall: [] thanksgiving: [] winter: [] interim: [] spring: [] easter: [] summer: - title: Hours hours: - {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'} <|file_sep|>updated/data/building-hours/3-2-print-center.yaml
name: Print Center image: print category: Mail and Packages schedule: - title: Hours # closedForChapelTime: true hours: # - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'} - {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'} - {days: [Fr], from: '8:00am', to: '12:00pm'} breakSchedule: fall: [] thanksgiving: [] winter: [] interim: [] spring: [] easter: [] summer: - title: Hours
<|file_sep|>data/building-hours/3-2-print-center.yaml.diff original: closedForChapelTime: true updated: # closedForChapelTime: true <|file_sep|>original/data/building-hours/3-2-print-center.yaml name: Print Center image: print category: Mail and Packages schedule: - title: Hours closedForChapelTime: true hours: # - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'} breakSchedule: fall: [] thanksgiving: [] winter: [] interim: [] spring: [] easter: [] summer: - title: Hours hours: - {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'} <|file_sep|>current/data/building-hours/3-2-print-center.yaml name: Print Center image: print category: Mail and Packages schedule: - title: Hours # closedForChapelTime: true hours: # - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'} breakSchedule: fall: [] thanksgiving: [] winter: [] interim: [] spring: [] easter: [] summer: - title: Hours hours: - {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'} <|file_sep|>updated/data/building-hours/3-2-print-center.yaml name: Print Center image: print category: Mail and Packages schedule: - title: Hours # closedForChapelTime: true hours: # - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'} - {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'} - {days: [Fr], from: '8:00am', to: '12:00pm'} breakSchedule: fall: [] thanksgiving: [] winter: [] interim: [] spring: [] easter: [] summer: - title: Hours
57c09ab37b8adcee99e375931404311acf0de878
data/building-hours/3-2-print-center.yaml
data/building-hours/3-2-print-center.yaml
YAML
<|file_sep|>original/generators/rspec/rspec_generator.rb def initialize(runtime_args, runtime_options = {}) Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks') super end def manifest record do |m| script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] } if Rails::VERSION::STRING < "2.1.0" m.file 'rspec.rake', 'lib/tasks/rspec.rake' end m.file 'script/autospec', 'script/autospec', script_options m.file 'script/spec', 'script/spec', script_options m.file 'script/spec_server', 'script/spec_server', script_options m.directory 'spec' m.file 'rcov.opts', 'spec/rcov.opts' m.file 'spec.opts', 'spec/spec.opts' m.template 'spec_helper.rb', 'spec/spec_helper.rb' <|file_sep|>current/generators/rspec/rspec_generator.rb def initialize(runtime_args, runtime_options = {}) Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks') super end def manifest record do |m| script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] } if Rails::VERSION::STRING < "2.1.0" m.file 'rspec.rake', 'lib/tasks/rspec.rake' end m.file 'script/autospec', 'script/autospec', script_options m.file 'script/spec', 'script/spec', script_options m.file 'script/spec_server', 'script/spec_server', script_options m.directory 'spec' m.file 'rcov.opts', 'spec/rcov.opts' m.file 'spec.opts', 'spec/spec.opts' m.template 'spec_helper.rb', 'spec/spec_helper.rb' <|file_sep|>updated/generators/rspec/rspec_generator.rb
def initialize(runtime_args, runtime_options = {}) Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks') super end def manifest record do |m| script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] } m.file 'rspec.rake', 'lib/tasks/rspec.rake' m.file 'script/autospec', 'script/autospec', script_options m.file 'script/spec', 'script/spec', script_options m.file 'script/spec_server', 'script/spec_server', script_options m.directory 'spec' m.file 'rcov.opts', 'spec/rcov.opts' m.file 'spec.opts', 'spec/spec.opts' m.template 'spec_helper.rb', 'spec/spec_helper.rb' m.directory 'stories'
<|file_sep|>original/generators/rspec/rspec_generator.rb def initialize(runtime_args, runtime_options = {}) Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks') super end def manifest record do |m| script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] } if Rails::VERSION::STRING < "2.1.0" m.file 'rspec.rake', 'lib/tasks/rspec.rake' end m.file 'script/autospec', 'script/autospec', script_options m.file 'script/spec', 'script/spec', script_options m.file 'script/spec_server', 'script/spec_server', script_options m.directory 'spec' m.file 'rcov.opts', 'spec/rcov.opts' m.file 'spec.opts', 'spec/spec.opts' m.template 'spec_helper.rb', 'spec/spec_helper.rb' <|file_sep|>current/generators/rspec/rspec_generator.rb def initialize(runtime_args, runtime_options = {}) Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks') super end def manifest record do |m| script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] } if Rails::VERSION::STRING < "2.1.0" m.file 'rspec.rake', 'lib/tasks/rspec.rake' end m.file 'script/autospec', 'script/autospec', script_options m.file 'script/spec', 'script/spec', script_options m.file 'script/spec_server', 'script/spec_server', script_options m.directory 'spec' m.file 'rcov.opts', 'spec/rcov.opts' m.file 'spec.opts', 'spec/spec.opts' m.template 'spec_helper.rb', 'spec/spec_helper.rb' <|file_sep|>updated/generators/rspec/rspec_generator.rb def initialize(runtime_args, runtime_options = {}) Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks') super end def manifest record do |m| script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] } m.file 'rspec.rake', 'lib/tasks/rspec.rake' m.file 'script/autospec', 'script/autospec', script_options m.file 'script/spec', 'script/spec', script_options m.file 'script/spec_server', 'script/spec_server', script_options m.directory 'spec' m.file 'rcov.opts', 'spec/rcov.opts' m.file 'spec.opts', 'spec/spec.opts' m.template 'spec_helper.rb', 'spec/spec_helper.rb' m.directory 'stories'
469315829dce46c37c337b260d67ca581189763f
generators/rspec/rspec_generator.rb
generators/rspec/rspec_generator.rb
Ruby
<|file_sep|>libs/utils.py.diff original: #get the cache key for storage updated: <|file_sep|>libs/utils.py.diff original: updated: """Get the cache key for storage""" <|file_sep|>libs/utils.py.diff original: #decorator for caching functions updated: <|file_sep|>libs/utils.py.diff original: updated: """Decorator for caching functions""" <|file_sep|>original/libs/utils.py serialise.append(str(arg)) for key,arg in kwargs.items(): if key == "clear_cache": continue serialise.append(str(key)) serialise.append(str(arg)) key = hashlib.md5("".join(serialise)).hexdigest() return key #decorator for caching functions def cache_for(time): def decorator(fn): def wrapper(*args, **kwargs): key = cache_get_key(fn.__name__, *args, **kwargs) result = cache.get(key) if not result or "clear_cache" in kwargs and kwargs["clear_cache"]: result = fn(*args, **kwargs) cache.set(key, result, time) return result return wrapper return decorator <|file_sep|>current/libs/utils.py for key,arg in kwargs.items(): if key == "clear_cache": continue serialise.append(str(key)) serialise.append(str(arg)) key = hashlib.md5("".join(serialise)).hexdigest() return key def cache_for(time): """Decorator for caching functions""" def decorator(fn): def wrapper(*args, **kwargs): key = cache_get_key(fn.__name__, *args, **kwargs) result = cache.get(key) if not result or "clear_cache" in kwargs and kwargs["clear_cache"]: result = fn(*args, **kwargs) cache.set(key, result, time) return result return wrapper return decorator <|file_sep|>updated/libs/utils.py
if key == "clear_cache": continue serialise.append(str(key)) serialise.append(str(arg)) key = hashlib.md5("".join(serialise)).hexdigest() return key def cache_for(time): """Decorator for caching functions""" def decorator(fn): def wrapper(*args, **kwargs): key = cache_get_key(fn.__name__, *args, **kwargs) result = cache.get(key) if not result or "clear_cache" in kwargs and kwargs["clear_cache"]: cache.delete(key) result = fn(*args, **kwargs) cache.set(key, result, time) return result return wrapper return decorator
<|file_sep|>libs/utils.py.diff original: #get the cache key for storage updated: <|file_sep|>libs/utils.py.diff original: updated: """Get the cache key for storage""" <|file_sep|>libs/utils.py.diff original: #decorator for caching functions updated: <|file_sep|>libs/utils.py.diff original: updated: """Decorator for caching functions""" <|file_sep|>original/libs/utils.py serialise.append(str(arg)) for key,arg in kwargs.items(): if key == "clear_cache": continue serialise.append(str(key)) serialise.append(str(arg)) key = hashlib.md5("".join(serialise)).hexdigest() return key #decorator for caching functions def cache_for(time): def decorator(fn): def wrapper(*args, **kwargs): key = cache_get_key(fn.__name__, *args, **kwargs) result = cache.get(key) if not result or "clear_cache" in kwargs and kwargs["clear_cache"]: result = fn(*args, **kwargs) cache.set(key, result, time) return result return wrapper return decorator <|file_sep|>current/libs/utils.py for key,arg in kwargs.items(): if key == "clear_cache": continue serialise.append(str(key)) serialise.append(str(arg)) key = hashlib.md5("".join(serialise)).hexdigest() return key def cache_for(time): """Decorator for caching functions""" def decorator(fn): def wrapper(*args, **kwargs): key = cache_get_key(fn.__name__, *args, **kwargs) result = cache.get(key) if not result or "clear_cache" in kwargs and kwargs["clear_cache"]: result = fn(*args, **kwargs) cache.set(key, result, time) return result return wrapper return decorator <|file_sep|>updated/libs/utils.py if key == "clear_cache": continue serialise.append(str(key)) serialise.append(str(arg)) key = hashlib.md5("".join(serialise)).hexdigest() return key def cache_for(time): """Decorator for caching functions""" def decorator(fn): def wrapper(*args, **kwargs): key = cache_get_key(fn.__name__, *args, **kwargs) result = cache.get(key) if not result or "clear_cache" in kwargs and kwargs["clear_cache"]: cache.delete(key) result = fn(*args, **kwargs) cache.set(key, result, time) return result return wrapper return decorator
709b9e57d8ea664715fd9bb89729f99324c3e0c2
libs/utils.py
libs/utils.py
Python
<|file_sep|>original/.travis.yml addons: firefox: latest google-chrome: latest install: - npm install - bower install before_script: - gulp lint version:check - polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html - xvfb-run -s '-screen 0 1024x768x24' wct script: - if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then wct --env saucelabs; fi - if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then wct --env saucelabs-cron; fi <|file_sep|>current/.travis.yml addons: firefox: latest google-chrome: latest install: - npm install - bower install before_script: - gulp lint version:check - polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html - xvfb-run -s '-screen 0 1024x768x24' wct script: - if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then wct --env saucelabs; fi - if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then wct --env saucelabs-cron; fi <|file_sep|>updated/.travis.yml
addons: firefox: latest google-chrome: latest install: - npm install - bower install before_script: - gulp lint version:check - polymer lint --rules polymer-2 --input ./src/*.html ./theme/**/*.html - xvfb-run -s '-screen 0 1024x768x24' wct script: - if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then wct --env saucelabs; fi - if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then wct --env saucelabs-cron; fi
<|file_sep|>original/.travis.yml addons: firefox: latest google-chrome: latest install: - npm install - bower install before_script: - gulp lint version:check - polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html - xvfb-run -s '-screen 0 1024x768x24' wct script: - if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then wct --env saucelabs; fi - if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then wct --env saucelabs-cron; fi <|file_sep|>current/.travis.yml addons: firefox: latest google-chrome: latest install: - npm install - bower install before_script: - gulp lint version:check - polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html - xvfb-run -s '-screen 0 1024x768x24' wct script: - if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then wct --env saucelabs; fi - if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then wct --env saucelabs-cron; fi <|file_sep|>updated/.travis.yml addons: firefox: latest google-chrome: latest install: - npm install - bower install before_script: - gulp lint version:check - polymer lint --rules polymer-2 --input ./src/*.html ./theme/**/*.html - xvfb-run -s '-screen 0 1024x768x24' wct script: - if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then wct --env saucelabs; fi - if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then wct --env saucelabs-cron; fi
cb748883f94154aac65822eca34966a8b4ca83e8
.travis.yml
.travis.yml
YAML
<|file_sep|>original/remove_symlinks.sh <|file_sep|>current/remove_symlinks.sh <|file_sep|>updated/remove_symlinks.sh
#!/bin/bash # -polka- # Get the current directory DIR="$( cd "$(dirname "$0" )" && pwd )" # Link dot files to $HOME for dotfile in $(find $DIR -maxdepth 1 -type f -name ".*"); do if [[ $(uname) == 'Linux' && "$(basename $dotfile)" == '.tmux.linux.conf' ]]; then rm -v "${HOME}/.tmux.conf" cp -v $dotfile "${HOME}/.tmux.conf" elif [[ $(uname) == 'Darwin' && "$(basename $dotfile)" == '.tmux.osx.conf' ]]; then rm -v "${HOME}/.tmux.conf" cp -v $dotfile "${HOME}/.tmux.conf" elif [[ "$(basename $dotfile)" != ".tmux.osx.conf" && "$(basename $dotfile)" != ".tmux.linux.conf" ]]; then rm -v "${HOME}/$(basename $dotfile)" cp -v $dotfile "${HOME}/$(basename $dotfile)" fi done # Link rc files in .bashrc.d
<|file_sep|>original/remove_symlinks.sh <|file_sep|>current/remove_symlinks.sh <|file_sep|>updated/remove_symlinks.sh #!/bin/bash # -polka- # Get the current directory DIR="$( cd "$(dirname "$0" )" && pwd )" # Link dot files to $HOME for dotfile in $(find $DIR -maxdepth 1 -type f -name ".*"); do if [[ $(uname) == 'Linux' && "$(basename $dotfile)" == '.tmux.linux.conf' ]]; then rm -v "${HOME}/.tmux.conf" cp -v $dotfile "${HOME}/.tmux.conf" elif [[ $(uname) == 'Darwin' && "$(basename $dotfile)" == '.tmux.osx.conf' ]]; then rm -v "${HOME}/.tmux.conf" cp -v $dotfile "${HOME}/.tmux.conf" elif [[ "$(basename $dotfile)" != ".tmux.osx.conf" && "$(basename $dotfile)" != ".tmux.linux.conf" ]]; then rm -v "${HOME}/$(basename $dotfile)" cp -v $dotfile "${HOME}/$(basename $dotfile)" fi done # Link rc files in .bashrc.d
8b15d01bb2c722bded40046c57fe052993d78ef4
remove_symlinks.sh
remove_symlinks.sh
Shell
<|file_sep|>original/DevMachineInstall/devApps.ps1 # Dev apps to install # Visual studio stuff cinstm VisualStudio2013Ultimate cinstm VisualStudio2012Ultimate # Visual studio add-ins cinstm xUnit.vs2012 cinstm WebEssentials2012 cinstm resharper -Version 7.1.3000.2254 cinstm typescript cinstm typescript.vs # databases and tools cinstm MsSqlServer2012Express cinstm mongodb cinstm MongoVUE # browsers cinstm Firefox cinstm GoogleChrome <|file_sep|>current/DevMachineInstall/devApps.ps1 # Dev apps to install # Visual studio stuff cinstm VisualStudio2013Ultimate cinstm VisualStudio2012Ultimate # Visual studio add-ins cinstm xUnit.vs2012 cinstm WebEssentials2012 cinstm resharper -Version 7.1.3000.2254 cinstm typescript cinstm typescript.vs # databases and tools cinstm MsSqlServer2012Express cinstm mongodb cinstm MongoVUE # browsers cinstm Firefox cinstm GoogleChrome <|file_sep|>updated/DevMachineInstall/devApps.ps1
# Dev apps to install # Visual studio stuff cinstm VisualStudio2013Ultimate cinstm VisualStudio2012Ultimate # Visual studio add-ins cinstm xUnit.vs2012 cinstm WebEssentials2012 cinstm resharper -Version 7.1.3000.2254 # Use only resharper if you want the latests. This is what I have a license for cinstm typescript cinstm typescript.vs # databases and tools cinstm MsSqlServer2012Express cinstm mongodb cinstm MongoVUE # browsers cinstm Firefox cinstm GoogleChrome
<|file_sep|>original/DevMachineInstall/devApps.ps1 # Dev apps to install # Visual studio stuff cinstm VisualStudio2013Ultimate cinstm VisualStudio2012Ultimate # Visual studio add-ins cinstm xUnit.vs2012 cinstm WebEssentials2012 cinstm resharper -Version 7.1.3000.2254 cinstm typescript cinstm typescript.vs # databases and tools cinstm MsSqlServer2012Express cinstm mongodb cinstm MongoVUE # browsers cinstm Firefox cinstm GoogleChrome <|file_sep|>current/DevMachineInstall/devApps.ps1 # Dev apps to install # Visual studio stuff cinstm VisualStudio2013Ultimate cinstm VisualStudio2012Ultimate # Visual studio add-ins cinstm xUnit.vs2012 cinstm WebEssentials2012 cinstm resharper -Version 7.1.3000.2254 cinstm typescript cinstm typescript.vs # databases and tools cinstm MsSqlServer2012Express cinstm mongodb cinstm MongoVUE # browsers cinstm Firefox cinstm GoogleChrome <|file_sep|>updated/DevMachineInstall/devApps.ps1 # Dev apps to install # Visual studio stuff cinstm VisualStudio2013Ultimate cinstm VisualStudio2012Ultimate # Visual studio add-ins cinstm xUnit.vs2012 cinstm WebEssentials2012 cinstm resharper -Version 7.1.3000.2254 # Use only resharper if you want the latests. This is what I have a license for cinstm typescript cinstm typescript.vs # databases and tools cinstm MsSqlServer2012Express cinstm mongodb cinstm MongoVUE # browsers cinstm Firefox cinstm GoogleChrome
2fd17380c9cac9b5fd69af3fddf1b01807cdbd0e
DevMachineInstall/devApps.ps1
DevMachineInstall/devApps.ps1
PowerShell
<|file_sep|>original/tests/units/FinderTest.php <|file_sep|>current/tests/units/FinderTest.php <|file_sep|>updated/tests/units/FinderTest.php
<?php use Ktp\Finder; use Ktp\Contracts\HttpClient as HttpClientContract; class FinderTest extends PHPUnit_Framework_TestCase { /** @test */ function finder_has_http_client() { $finder = new Finder; $this->assertInstanceOf(HttpClientContract::class, $finder->httpClient()); } /** @test */ function finder_find_return_null_when_nik_is_not_found() { $finder = new Finder; $this->assertNull($finder->findByNik(123)); }
<|file_sep|>original/tests/units/FinderTest.php <|file_sep|>current/tests/units/FinderTest.php <|file_sep|>updated/tests/units/FinderTest.php <?php use Ktp\Finder; use Ktp\Contracts\HttpClient as HttpClientContract; class FinderTest extends PHPUnit_Framework_TestCase { /** @test */ function finder_has_http_client() { $finder = new Finder; $this->assertInstanceOf(HttpClientContract::class, $finder->httpClient()); } /** @test */ function finder_find_return_null_when_nik_is_not_found() { $finder = new Finder; $this->assertNull($finder->findByNik(123)); }
96cb061f49e469957c440a2dae5a01bf6bda39fd
tests/units/FinderTest.php
tests/units/FinderTest.php
PHP
<|file_sep|>original/README.md #### TaskRunner / Building * Grunt * Webpack * Babel #### Testing * Mocha * Sinon * Chai #### Backend + DB * Node.js * Express * Mongoose * MongoDB #### Map / Geolocation * Leaflet.js #### Graphic Icon Front Lawn House by Alfredo Hernandez from the Noun Project <|file_sep|>current/README.md #### TaskRunner / Building * Grunt * Webpack * Babel #### Testing * Mocha * Sinon * Chai #### Backend + DB * Node.js * Express * Mongoose * MongoDB #### Map / Geolocation * Leaflet.js #### Graphic Icon Front Lawn House by Alfredo Hernandez from the Noun Project <|file_sep|>updated/README.md
#### TaskRunner / Building * Grunt * Webpack * Babel #### Testing * Mocha * Sinon * Chai #### Backend + DB * Node.js * Express * Mongoose * MongoDB #### Map / Geolocation * Leaflet.js #### Graphic Icon * House by Vishwas Shetty from the Noun Project
<|file_sep|>original/README.md #### TaskRunner / Building * Grunt * Webpack * Babel #### Testing * Mocha * Sinon * Chai #### Backend + DB * Node.js * Express * Mongoose * MongoDB #### Map / Geolocation * Leaflet.js #### Graphic Icon Front Lawn House by Alfredo Hernandez from the Noun Project <|file_sep|>current/README.md #### TaskRunner / Building * Grunt * Webpack * Babel #### Testing * Mocha * Sinon * Chai #### Backend + DB * Node.js * Express * Mongoose * MongoDB #### Map / Geolocation * Leaflet.js #### Graphic Icon Front Lawn House by Alfredo Hernandez from the Noun Project <|file_sep|>updated/README.md #### TaskRunner / Building * Grunt * Webpack * Babel #### Testing * Mocha * Sinon * Chai #### Backend + DB * Node.js * Express * Mongoose * MongoDB #### Map / Geolocation * Leaflet.js #### Graphic Icon * House by Vishwas Shetty from the Noun Project
89858de26d73818bf48ff9915087cdea34a3fd21
README.md
README.md
Markdown
<|file_sep|>original/PATREON.md Patreon Support --------------- The Espruino Web IDE (and Espruino in general) is supported by generous donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a> Supporters that pay $10/month or more to help support Espruino are: Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher, Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull <|file_sep|>current/PATREON.md Patreon Support --------------- The Espruino Web IDE (and Espruino in general) is supported by generous donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a> Supporters that pay $10/month or more to help support Espruino are: Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher, Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull <|file_sep|>updated/PATREON.md
Patreon Support --------------- The Espruino Web IDE (and Espruino in general) is supported by generous donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a> Supporters that pay $10/month or more to help support Espruino are: Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher, Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull, Gito Nirmolo
<|file_sep|>original/PATREON.md Patreon Support --------------- The Espruino Web IDE (and Espruino in general) is supported by generous donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a> Supporters that pay $10/month or more to help support Espruino are: Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher, Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull <|file_sep|>current/PATREON.md Patreon Support --------------- The Espruino Web IDE (and Espruino in general) is supported by generous donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a> Supporters that pay $10/month or more to help support Espruino are: Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher, Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull <|file_sep|>updated/PATREON.md Patreon Support --------------- The Espruino Web IDE (and Espruino in general) is supported by generous donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a> Supporters that pay $10/month or more to help support Espruino are: Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher, Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull, Gito Nirmolo
33af531a529df7382a32aeb79d1057808239a870
PATREON.md
PATREON.md
Markdown
<|file_sep|>app/views/questions/_votes.html.erb.diff original: <%= form_for question, url: questions_upvote_path, method: :post do |f| %> updated: <%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %> <|file_sep|>original/app/views/questions/_votes.html.erb <div class="vote"> <%= form_for question, url: questions_upvote_path, method: :post do |f| %> <input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <%= form_for question, url: questions_downvote_path, method: :post do |f| %> <input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <span class="votes_count"> <%= question.votes_count %> </span> </div> <|file_sep|>current/app/views/questions/_votes.html.erb <div class="vote"> <%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %> <input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <%= form_for question, url: questions_downvote_path, method: :post do |f| %> <input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <span class="votes_count"> <%= question.votes_count %> </span> </div> <|file_sep|>updated/app/views/questions/_votes.html.erb
<div class="vote"> <%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %> <input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <%= form_for question, url: questions_downvote_path, method: :post, html: {class: "downvote_question"} do |f| %> <input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <span class="votes_count"> <%= question.votes_count %> </span> </div>
<|file_sep|>app/views/questions/_votes.html.erb.diff original: <%= form_for question, url: questions_upvote_path, method: :post do |f| %> updated: <%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %> <|file_sep|>original/app/views/questions/_votes.html.erb <div class="vote"> <%= form_for question, url: questions_upvote_path, method: :post do |f| %> <input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <%= form_for question, url: questions_downvote_path, method: :post do |f| %> <input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <span class="votes_count"> <%= question.votes_count %> </span> </div> <|file_sep|>current/app/views/questions/_votes.html.erb <div class="vote"> <%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %> <input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <%= form_for question, url: questions_downvote_path, method: :post do |f| %> <input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <span class="votes_count"> <%= question.votes_count %> </span> </div> <|file_sep|>updated/app/views/questions/_votes.html.erb <div class="vote"> <%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %> <input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <%= form_for question, url: questions_downvote_path, method: :post, html: {class: "downvote_question"} do |f| %> <input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/> <input type="hidden" name="id" value="<%= question.id %>"/> <% end %> <span class="votes_count"> <%= question.votes_count %> </span> </div>
2edd043877ba222f7ad0761761a9379a660378cd
app/views/questions/_votes.html.erb
app/views/questions/_votes.html.erb
HTML+ERB
<|file_sep|>README.md.diff original: updated: [![Build Status](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror.svg?branch=develop)](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror) [![Requirements Status](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements.svg?branch=develop)](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop) <|file_sep|>original/README.md # debian-wheel-mirror Automated Python wheel building and deployment for https://pypi.p16n.org This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles). These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org). The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`. <|file_sep|>current/README.md # debian-wheel-mirror [![Build Status](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror.svg?branch=develop)](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror) [![Requirements Status](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements.svg?branch=develop)](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop) Automated Python wheel building and deployment for https://pypi.p16n.org This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles). These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org). The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`. <|file_sep|>updated/README.md
# debian-wheel-mirror [![Build Status](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror.svg?branch=develop)](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror) [![Requirements Status](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements.svg?branch=develop)](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop) Automated Python wheel building and deployment for https://pypi.p16n.org This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles). The packages are kept up-to-date by automated pull requests made by [Requires.io](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop). Packages are built inside Docker containers (currently using the [`python:2`](https://hub.docker.com/_/python/) image) and uploaded to our PyPi mirror by [Travis CI](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror). These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org). The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`.
<|file_sep|>README.md.diff original: updated: [![Build Status](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror.svg?branch=develop)](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror) [![Requirements Status](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements.svg?branch=develop)](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop) <|file_sep|>original/README.md # debian-wheel-mirror Automated Python wheel building and deployment for https://pypi.p16n.org This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles). These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org). The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`. <|file_sep|>current/README.md # debian-wheel-mirror [![Build Status](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror.svg?branch=develop)](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror) [![Requirements Status](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements.svg?branch=develop)](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop) Automated Python wheel building and deployment for https://pypi.p16n.org This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles). These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org). The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`. <|file_sep|>updated/README.md # debian-wheel-mirror [![Build Status](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror.svg?branch=develop)](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror) [![Requirements Status](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements.svg?branch=develop)](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop) Automated Python wheel building and deployment for https://pypi.p16n.org This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles). The packages are kept up-to-date by automated pull requests made by [Requires.io](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop). Packages are built inside Docker containers (currently using the [`python:2`](https://hub.docker.com/_/python/) image) and uploaded to our PyPi mirror by [Travis CI](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror). These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org). The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`.
3fa729e7fb11f05173ea14b0fed3294aff715c71
README.md
README.md
Markdown
<|file_sep|>benchmark-perf.gemspec.diff original: # coding: utf-8 updated: <|file_sep|>original/benchmark-perf.gemspec lib = File.expand_path('../lib', __FILE__) $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib) require 'benchmark/perf/version' Gem::Specification.new do |spec| spec.name = "benchmark-perf" spec.version = Benchmark::Perf::VERSION spec.authors = ["Piotr Murach"] spec.email = [""] spec.summary = %q{Execution time and iteration performance benchmarking} spec.description = %q{Execution time and iteration performance benchmarking} spec.homepage = "" spec.license = "MIT" spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^spec/}) spec.require_paths = ["lib"] spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0' spec.add_development_dependency 'rake' <|file_sep|>current/benchmark-perf.gemspec $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib) require 'benchmark/perf/version' Gem::Specification.new do |spec| spec.name = "benchmark-perf" spec.version = Benchmark::Perf::VERSION spec.authors = ["Piotr Murach"] spec.email = [""] spec.summary = %q{Execution time and iteration performance benchmarking} spec.description = %q{Execution time and iteration performance benchmarking} spec.homepage = "" spec.license = "MIT" spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^spec/}) spec.require_paths = ["lib"] spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0' spec.add_development_dependency 'rake' end <|file_sep|>updated/benchmark-perf.gemspec
require 'benchmark/perf/version' Gem::Specification.new do |spec| spec.name = "benchmark-perf" spec.version = Benchmark::Perf::VERSION spec.authors = ["Piotr Murach"] spec.email = [""] spec.summary = %q{Execution time and iteration performance benchmarking} spec.description = %q{Execution time and iteration performance benchmarking} spec.homepage = "" spec.license = "MIT" spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^spec/}) spec.require_paths = ["lib"] spec.add_development_dependency 'bundler', '~> 1.16' spec.add_development_dependency 'rspec', '~> 3.0' spec.add_development_dependency 'rake', '~> 10.0' end
<|file_sep|>benchmark-perf.gemspec.diff original: # coding: utf-8 updated: <|file_sep|>original/benchmark-perf.gemspec lib = File.expand_path('../lib', __FILE__) $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib) require 'benchmark/perf/version' Gem::Specification.new do |spec| spec.name = "benchmark-perf" spec.version = Benchmark::Perf::VERSION spec.authors = ["Piotr Murach"] spec.email = [""] spec.summary = %q{Execution time and iteration performance benchmarking} spec.description = %q{Execution time and iteration performance benchmarking} spec.homepage = "" spec.license = "MIT" spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^spec/}) spec.require_paths = ["lib"] spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0' spec.add_development_dependency 'rake' <|file_sep|>current/benchmark-perf.gemspec $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib) require 'benchmark/perf/version' Gem::Specification.new do |spec| spec.name = "benchmark-perf" spec.version = Benchmark::Perf::VERSION spec.authors = ["Piotr Murach"] spec.email = [""] spec.summary = %q{Execution time and iteration performance benchmarking} spec.description = %q{Execution time and iteration performance benchmarking} spec.homepage = "" spec.license = "MIT" spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^spec/}) spec.require_paths = ["lib"] spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0' spec.add_development_dependency 'rake' end <|file_sep|>updated/benchmark-perf.gemspec require 'benchmark/perf/version' Gem::Specification.new do |spec| spec.name = "benchmark-perf" spec.version = Benchmark::Perf::VERSION spec.authors = ["Piotr Murach"] spec.email = [""] spec.summary = %q{Execution time and iteration performance benchmarking} spec.description = %q{Execution time and iteration performance benchmarking} spec.homepage = "" spec.license = "MIT" spec.files = `git ls-files -z`.split("\x0") spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) } spec.test_files = spec.files.grep(%r{^spec/}) spec.require_paths = ["lib"] spec.add_development_dependency 'bundler', '~> 1.16' spec.add_development_dependency 'rspec', '~> 3.0' spec.add_development_dependency 'rake', '~> 10.0' end
d2f12484c620c7056d028fa552ee56a03233bf71
benchmark-perf.gemspec
benchmark-perf.gemspec
Ruby
<|file_sep|>.travis.yml.diff original: updated: - "npm install -g markdownlint-cli" <|file_sep|>original/.travis.yml cache: "pip" before_script: # Run linters - "flake8 ." - "prospector --strictness veryhigh" - "pylint src/vws" - "mypy src tests" # Make sure that imports are sorted as expected. - "isort --recursive --check-only" # Create an environment file used by `pytest-envfiles`. # We don't add to this because we have environment variables set in the CI # config. - "touch vuforia_secrets.env" script: - "pytest --cov=src --cov=tests" after_success: - "coveralls" <|file_sep|>current/.travis.yml - "npm install -g markdownlint-cli" cache: "pip" before_script: # Run linters - "flake8 ." - "prospector --strictness veryhigh" - "pylint src/vws" - "mypy src tests" # Make sure that imports are sorted as expected. - "isort --recursive --check-only" # Create an environment file used by `pytest-envfiles`. # We don't add to this because we have environment variables set in the CI # config. - "touch vuforia_secrets.env" script: - "pytest --cov=src --cov=tests" after_success: <|file_sep|>updated/.travis.yml
- "npm install -g markdownlint-cli" cache: "pip" before_script: # Run linters - "flake8 ." - "prospector --strictness veryhigh" - "pylint src/vws" - "mypy src tests" - "markdownlint README.md" # Make sure that imports are sorted as expected. - "isort --recursive --check-only" # Create an environment file used by `pytest-envfiles`. # We don't add to this because we have environment variables set in the CI # config. - "touch vuforia_secrets.env" script: - "pytest --cov=src --cov=tests"
<|file_sep|>.travis.yml.diff original: updated: - "npm install -g markdownlint-cli" <|file_sep|>original/.travis.yml cache: "pip" before_script: # Run linters - "flake8 ." - "prospector --strictness veryhigh" - "pylint src/vws" - "mypy src tests" # Make sure that imports are sorted as expected. - "isort --recursive --check-only" # Create an environment file used by `pytest-envfiles`. # We don't add to this because we have environment variables set in the CI # config. - "touch vuforia_secrets.env" script: - "pytest --cov=src --cov=tests" after_success: - "coveralls" <|file_sep|>current/.travis.yml - "npm install -g markdownlint-cli" cache: "pip" before_script: # Run linters - "flake8 ." - "prospector --strictness veryhigh" - "pylint src/vws" - "mypy src tests" # Make sure that imports are sorted as expected. - "isort --recursive --check-only" # Create an environment file used by `pytest-envfiles`. # We don't add to this because we have environment variables set in the CI # config. - "touch vuforia_secrets.env" script: - "pytest --cov=src --cov=tests" after_success: <|file_sep|>updated/.travis.yml - "npm install -g markdownlint-cli" cache: "pip" before_script: # Run linters - "flake8 ." - "prospector --strictness veryhigh" - "pylint src/vws" - "mypy src tests" - "markdownlint README.md" # Make sure that imports are sorted as expected. - "isort --recursive --check-only" # Create an environment file used by `pytest-envfiles`. # We don't add to this because we have environment variables set in the CI # config. - "touch vuforia_secrets.env" script: - "pytest --cov=src --cov=tests"
93e3ecf683d789df795c9bd1d7186268916ceeb0
.travis.yml
.travis.yml
YAML
<|file_sep|>packages/mw/mwc-probability-transition.yaml.diff original: hash: 42d82dd5cb90933637d7e247e88f0935fc75f8ea3e968e6abd1cec469ba9be8e updated: hash: abeae477b2f800f78cf4f4f9fc86b103472cdde5b2ef4761abfecbf2076b0e08 <|file_sep|>packages/mw/mwc-probability-transition.yaml.diff original: updated: - '0.3.0.1' <|file_sep|>original/packages/mw/mwc-probability-transition.yaml mtl: -any transformers: -any mwc-probability: -any primitive: -any all-versions: - '0.1.0.0' - '0.2.0.0' - '0.3.0.0' author: Marco Zocca latest: '0.3.0.0' description-type: markdown description: ! '# mwc-probability-transition [![Build Status](https://travis-ci.org/ocramz/mwc-probability-transition.png)](https://travis-ci.org/ocramz/mwc-probability-transition) Types and primitives for stochastic simulation (e.g. integration of SDE, random walks, Markov Chain Monte Carlo algorithms etc.) ' <|file_sep|>current/packages/mw/mwc-probability-transition.yaml mtl: -any transformers: -any mwc-probability: -any primitive: -any all-versions: - '0.1.0.0' - '0.2.0.0' - '0.3.0.0' - '0.3.0.1' author: Marco Zocca latest: '0.3.0.0' description-type: markdown description: ! '# mwc-probability-transition [![Build Status](https://travis-ci.org/ocramz/mwc-probability-transition.png)](https://travis-ci.org/ocramz/mwc-probability-transition) Types and primitives for stochastic simulation (e.g. integration of SDE, random walks, Markov Chain Monte Carlo algorithms etc.) <|file_sep|>updated/packages/mw/mwc-probability-transition.yaml
mtl: -any transformers: -any mwc-probability: -any primitive: -any all-versions: - '0.1.0.0' - '0.2.0.0' - '0.3.0.0' - '0.3.0.1' author: Marco Zocca latest: '0.3.0.1' description-type: markdown description: ! '# mwc-probability-transition [![Build Status](https://travis-ci.org/ocramz/mwc-probability-transition.png)](https://travis-ci.org/ocramz/mwc-probability-transition) Types and primitives for stochastic simulation (e.g. integration of SDE, random walks, Markov Chain Monte Carlo algorithms etc.)
<|file_sep|>packages/mw/mwc-probability-transition.yaml.diff original: hash: 42d82dd5cb90933637d7e247e88f0935fc75f8ea3e968e6abd1cec469ba9be8e updated: hash: abeae477b2f800f78cf4f4f9fc86b103472cdde5b2ef4761abfecbf2076b0e08 <|file_sep|>packages/mw/mwc-probability-transition.yaml.diff original: updated: - '0.3.0.1' <|file_sep|>original/packages/mw/mwc-probability-transition.yaml mtl: -any transformers: -any mwc-probability: -any primitive: -any all-versions: - '0.1.0.0' - '0.2.0.0' - '0.3.0.0' author: Marco Zocca latest: '0.3.0.0' description-type: markdown description: ! '# mwc-probability-transition [![Build Status](https://travis-ci.org/ocramz/mwc-probability-transition.png)](https://travis-ci.org/ocramz/mwc-probability-transition) Types and primitives for stochastic simulation (e.g. integration of SDE, random walks, Markov Chain Monte Carlo algorithms etc.) ' <|file_sep|>current/packages/mw/mwc-probability-transition.yaml mtl: -any transformers: -any mwc-probability: -any primitive: -any all-versions: - '0.1.0.0' - '0.2.0.0' - '0.3.0.0' - '0.3.0.1' author: Marco Zocca latest: '0.3.0.0' description-type: markdown description: ! '# mwc-probability-transition [![Build Status](https://travis-ci.org/ocramz/mwc-probability-transition.png)](https://travis-ci.org/ocramz/mwc-probability-transition) Types and primitives for stochastic simulation (e.g. integration of SDE, random walks, Markov Chain Monte Carlo algorithms etc.) <|file_sep|>updated/packages/mw/mwc-probability-transition.yaml mtl: -any transformers: -any mwc-probability: -any primitive: -any all-versions: - '0.1.0.0' - '0.2.0.0' - '0.3.0.0' - '0.3.0.1' author: Marco Zocca latest: '0.3.0.1' description-type: markdown description: ! '# mwc-probability-transition [![Build Status](https://travis-ci.org/ocramz/mwc-probability-transition.png)](https://travis-ci.org/ocramz/mwc-probability-transition) Types and primitives for stochastic simulation (e.g. integration of SDE, random walks, Markov Chain Monte Carlo algorithms etc.)
bae4265c31082a932a8c4b3dfaa70bbab4635bdd
packages/mw/mwc-probability-transition.yaml
packages/mw/mwc-probability-transition.yaml
YAML
<|file_sep|>original/server/requirements.txt gunicorn==19.4.5 honcho==0.6.6 Eve==0.6.3 newrelic>=2.66,<2.67 -e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core <|file_sep|>current/server/requirements.txt gunicorn==19.4.5 honcho==0.6.6 Eve==0.6.3 newrelic>=2.66,<2.67 -e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core <|file_sep|>updated/server/requirements.txt
gunicorn==19.4.5 honcho==0.6.6 Eve==0.6.3 newrelic>=2.66,<2.67 -e git+git://github.com/superdesk/superdesk-core@7fca944#egg=Superdesk-Core
<|file_sep|>original/server/requirements.txt gunicorn==19.4.5 honcho==0.6.6 Eve==0.6.3 newrelic>=2.66,<2.67 -e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core <|file_sep|>current/server/requirements.txt gunicorn==19.4.5 honcho==0.6.6 Eve==0.6.3 newrelic>=2.66,<2.67 -e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core <|file_sep|>updated/server/requirements.txt gunicorn==19.4.5 honcho==0.6.6 Eve==0.6.3 newrelic>=2.66,<2.67 -e git+git://github.com/superdesk/superdesk-core@7fca944#egg=Superdesk-Core
6bafeb289f56ddc41360384a42c0fa67c63ba5dd
server/requirements.txt
server/requirements.txt
Text
<|file_sep|>original/iobuf/ibuf_readall.c #include <iobuf/iobuf.h> #include <str/str.h> /** Read the remainder of the \c ibuf into the \c str. */ int ibuf_readall(ibuf* in, str* out) { for (;;) { if (!str_catb(out, in->io.buffer+in->io.bufstart, in->io.buflen-in->io.bufstart)) return 0; in->io.bufstart = in->io.buflen; if (!ibuf_refill(in)) return ibuf_eof(in); } } <|file_sep|>current/iobuf/ibuf_readall.c #include <iobuf/iobuf.h> #include <str/str.h> /** Read the remainder of the \c ibuf into the \c str. */ int ibuf_readall(ibuf* in, str* out) { for (;;) { if (!str_catb(out, in->io.buffer+in->io.bufstart, in->io.buflen-in->io.bufstart)) return 0; in->io.bufstart = in->io.buflen; if (!ibuf_refill(in)) return ibuf_eof(in); } } <|file_sep|>updated/iobuf/ibuf_readall.c
#include <iobuf/iobuf.h> #include <str/str.h> /** Read the remainder of the \c ibuf into the \c str. */ int ibuf_readall(ibuf* in, str* out) { if (ibuf_eof(in)) return 1; if (ibuf_error(in)) return 0; for (;;) { if (!str_catb(out, in->io.buffer+in->io.bufstart, in->io.buflen-in->io.bufstart)) return 0; in->io.bufstart = in->io.buflen; if (!ibuf_refill(in)) return ibuf_eof(in); } }
<|file_sep|>original/iobuf/ibuf_readall.c #include <iobuf/iobuf.h> #include <str/str.h> /** Read the remainder of the \c ibuf into the \c str. */ int ibuf_readall(ibuf* in, str* out) { for (;;) { if (!str_catb(out, in->io.buffer+in->io.bufstart, in->io.buflen-in->io.bufstart)) return 0; in->io.bufstart = in->io.buflen; if (!ibuf_refill(in)) return ibuf_eof(in); } } <|file_sep|>current/iobuf/ibuf_readall.c #include <iobuf/iobuf.h> #include <str/str.h> /** Read the remainder of the \c ibuf into the \c str. */ int ibuf_readall(ibuf* in, str* out) { for (;;) { if (!str_catb(out, in->io.buffer+in->io.bufstart, in->io.buflen-in->io.bufstart)) return 0; in->io.bufstart = in->io.buflen; if (!ibuf_refill(in)) return ibuf_eof(in); } } <|file_sep|>updated/iobuf/ibuf_readall.c #include <iobuf/iobuf.h> #include <str/str.h> /** Read the remainder of the \c ibuf into the \c str. */ int ibuf_readall(ibuf* in, str* out) { if (ibuf_eof(in)) return 1; if (ibuf_error(in)) return 0; for (;;) { if (!str_catb(out, in->io.buffer+in->io.bufstart, in->io.buflen-in->io.bufstart)) return 0; in->io.bufstart = in->io.buflen; if (!ibuf_refill(in)) return ibuf_eof(in); } }
17530c3d7eceb106a74066446404117158e8aa93
iobuf/ibuf_readall.c
iobuf/ibuf_readall.c
C
<|file_sep|>original/.travis.yml sudo: true language: cpp os: - osx compiler: - gcc cache: directories: - "$HOME/Library/Caches/Homebrew" script: - curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O - chmod u+x macpython-download-cache-and-build-module-wheels.sh - ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5 - tar -zcvf dist.tar.gz dist/ - curl -F file="@dist.tar.gz" https://file.io <|file_sep|>current/.travis.yml sudo: true language: cpp os: - osx compiler: - gcc cache: directories: - "$HOME/Library/Caches/Homebrew" script: - curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O - chmod u+x macpython-download-cache-and-build-module-wheels.sh - ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5 - tar -zcvf dist.tar.gz dist/ - curl -F file="@dist.tar.gz" https://file.io <|file_sep|>updated/.travis.yml
sudo: true language: cpp os: - osx compiler: - gcc cache: directories: - "$HOME/Library/Caches/Homebrew" script: - curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O - chmod u+x macpython-download-cache-and-build-module-wheels.sh - ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5 - tar -zcvf dist.tar.gz dist/ - curl --upload-file dist.tar.gz https://transfer.sh/dist.tar.gz
<|file_sep|>original/.travis.yml sudo: true language: cpp os: - osx compiler: - gcc cache: directories: - "$HOME/Library/Caches/Homebrew" script: - curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O - chmod u+x macpython-download-cache-and-build-module-wheels.sh - ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5 - tar -zcvf dist.tar.gz dist/ - curl -F file="@dist.tar.gz" https://file.io <|file_sep|>current/.travis.yml sudo: true language: cpp os: - osx compiler: - gcc cache: directories: - "$HOME/Library/Caches/Homebrew" script: - curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O - chmod u+x macpython-download-cache-and-build-module-wheels.sh - ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5 - tar -zcvf dist.tar.gz dist/ - curl -F file="@dist.tar.gz" https://file.io <|file_sep|>updated/.travis.yml sudo: true language: cpp os: - osx compiler: - gcc cache: directories: - "$HOME/Library/Caches/Homebrew" script: - curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O - chmod u+x macpython-download-cache-and-build-module-wheels.sh - ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5 - tar -zcvf dist.tar.gz dist/ - curl --upload-file dist.tar.gz https://transfer.sh/dist.tar.gz
4757fb4442dfb1b1b4444a281fc3b246b0952cb7
.travis.yml
.travis.yml
YAML
<|file_sep|>original/spec/models/user_spec.rb <|file_sep|>current/spec/models/user_spec.rb <|file_sep|>updated/spec/models/user_spec.rb
require 'rails_helper' require 'factory_girl_rails' describe User do let(:user){ FactoryGirl.create(:user) } it "has a valid FACTORY" do expect(FactoryGirl.create(:user)).to be_valid end it "should have an email address" do #when left blank, this tests for truthiness expect(user.email).to be end it "should have a valid email address" do expect(user.email).to match(/^\w+.\w+@\w+.\w+$/) end it 'should be able to create snippets' do expect(user.snippets).to be
<|file_sep|>original/spec/models/user_spec.rb <|file_sep|>current/spec/models/user_spec.rb <|file_sep|>updated/spec/models/user_spec.rb require 'rails_helper' require 'factory_girl_rails' describe User do let(:user){ FactoryGirl.create(:user) } it "has a valid FACTORY" do expect(FactoryGirl.create(:user)).to be_valid end it "should have an email address" do #when left blank, this tests for truthiness expect(user.email).to be end it "should have a valid email address" do expect(user.email).to match(/^\w+.\w+@\w+.\w+$/) end it 'should be able to create snippets' do expect(user.snippets).to be
a1a1830a4cee6d50d7a58b2ddf82105e9c5d255b
spec/models/user_spec.rb
spec/models/user_spec.rb
Ruby
<|file_sep|>README.md.diff original: updated: ## Getting going Here's how to give this a try. First, you'll need Postgres installed and running. Next, run these commands: ```bash $ git clone https://github.com/oestrich/hypermedia_rails.git $ cd hypermedia_rails $ bundle $ RAILS_ENV=test bundle exec rake db:create db:migrate $ RAILS_ENV=development bundle exec rake db:create db:migrate $ bundle exec rake spec ``` You should see all the tests pass. <|file_sep|>original/README.md # Hypermedia API with Rails This is a Hypermedia API example in Rails. It uses several cool gems that are good for hypermedia and/or APIs: * active_model_serializers * rspec_api_documentation * raddocs ## Hypermedia client A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token. <|file_sep|>current/README.md * raddocs ## Getting going Here's how to give this a try. First, you'll need Postgres installed and running. Next, run these commands: ```bash $ git clone https://github.com/oestrich/hypermedia_rails.git $ cd hypermedia_rails $ bundle $ RAILS_ENV=test bundle exec rake db:create db:migrate $ RAILS_ENV=development bundle exec rake db:create db:migrate $ bundle exec rake spec ``` You should see all the tests pass. ## Hypermedia client A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token. <|file_sep|>updated/README.md
```bash $ git clone https://github.com/oestrich/hypermedia_rails.git $ cd hypermedia_rails $ bundle $ RAILS_ENV=test bundle exec rake db:create db:migrate $ RAILS_ENV=development bundle exec rake db:create db:migrate $ bundle exec rake spec ``` You should see all the tests pass. ## Hypermedia client A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token. To try it, just ```bash $ bundle exec ruby hypermedia.rb ```
<|file_sep|>README.md.diff original: updated: ## Getting going Here's how to give this a try. First, you'll need Postgres installed and running. Next, run these commands: ```bash $ git clone https://github.com/oestrich/hypermedia_rails.git $ cd hypermedia_rails $ bundle $ RAILS_ENV=test bundle exec rake db:create db:migrate $ RAILS_ENV=development bundle exec rake db:create db:migrate $ bundle exec rake spec ``` You should see all the tests pass. <|file_sep|>original/README.md # Hypermedia API with Rails This is a Hypermedia API example in Rails. It uses several cool gems that are good for hypermedia and/or APIs: * active_model_serializers * rspec_api_documentation * raddocs ## Hypermedia client A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token. <|file_sep|>current/README.md * raddocs ## Getting going Here's how to give this a try. First, you'll need Postgres installed and running. Next, run these commands: ```bash $ git clone https://github.com/oestrich/hypermedia_rails.git $ cd hypermedia_rails $ bundle $ RAILS_ENV=test bundle exec rake db:create db:migrate $ RAILS_ENV=development bundle exec rake db:create db:migrate $ bundle exec rake spec ``` You should see all the tests pass. ## Hypermedia client A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token. <|file_sep|>updated/README.md ```bash $ git clone https://github.com/oestrich/hypermedia_rails.git $ cd hypermedia_rails $ bundle $ RAILS_ENV=test bundle exec rake db:create db:migrate $ RAILS_ENV=development bundle exec rake db:create db:migrate $ bundle exec rake spec ``` You should see all the tests pass. ## Hypermedia client A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token. To try it, just ```bash $ bundle exec ruby hypermedia.rb ```
82ae82332ff37ffdbc17a4d1ae420a01ba4d88cc
README.md
README.md
Markdown
<|file_sep|>original/composer.json { "name": "nexylan/paybox-direct-plus", "description": "PayBox Direct Plus API wrapper", "type": "library", "license": "MIT", "authors": [ { "name": "Sullivan SENECHAL", "email": "soullivaneuh@gmail.com" } ], "require": { "php": "^5.5 || ^7.0", "guzzlehttp/guzzle": "^6.0", "symfony/options-resolver": "^2.7 || ^3.0" }, "require-dev": { "phpunit/phpunit": "^5.3", "sllh/php-cs-fixer-styleci-bridge": "^2.0" }, "autoload": { <|file_sep|>current/composer.json { "name": "nexylan/paybox-direct-plus", "description": "PayBox Direct Plus API wrapper", "type": "library", "license": "MIT", "authors": [ { "name": "Sullivan SENECHAL", "email": "soullivaneuh@gmail.com" } ], "require": { "php": "^5.5 || ^7.0", "guzzlehttp/guzzle": "^6.0", "symfony/options-resolver": "^2.7 || ^3.0" }, "require-dev": { "phpunit/phpunit": "^5.3", "sllh/php-cs-fixer-styleci-bridge": "^2.0" }, "autoload": { <|file_sep|>updated/composer.json
{ "name": "nexylan/paybox-direct", "description": "PayBox Direct and Direct Plus API wrapper", "keywords": ["Paybox", "Direct", "API", "Symfony"], "type": "library", "license": "MIT", "authors": [ { "name": "Sullivan SENECHAL", "email": "soullivaneuh@gmail.com" } ], "require": { "php": "^5.5 || ^7.0", "guzzlehttp/guzzle": "^6.0", "symfony/options-resolver": "^2.7 || ^3.0" }, "require-dev": { "phpunit/phpunit": "^5.3", "sllh/php-cs-fixer-styleci-bridge": "^2.0" },
<|file_sep|>original/composer.json { "name": "nexylan/paybox-direct-plus", "description": "PayBox Direct Plus API wrapper", "type": "library", "license": "MIT", "authors": [ { "name": "Sullivan SENECHAL", "email": "soullivaneuh@gmail.com" } ], "require": { "php": "^5.5 || ^7.0", "guzzlehttp/guzzle": "^6.0", "symfony/options-resolver": "^2.7 || ^3.0" }, "require-dev": { "phpunit/phpunit": "^5.3", "sllh/php-cs-fixer-styleci-bridge": "^2.0" }, "autoload": { <|file_sep|>current/composer.json { "name": "nexylan/paybox-direct-plus", "description": "PayBox Direct Plus API wrapper", "type": "library", "license": "MIT", "authors": [ { "name": "Sullivan SENECHAL", "email": "soullivaneuh@gmail.com" } ], "require": { "php": "^5.5 || ^7.0", "guzzlehttp/guzzle": "^6.0", "symfony/options-resolver": "^2.7 || ^3.0" }, "require-dev": { "phpunit/phpunit": "^5.3", "sllh/php-cs-fixer-styleci-bridge": "^2.0" }, "autoload": { <|file_sep|>updated/composer.json { "name": "nexylan/paybox-direct", "description": "PayBox Direct and Direct Plus API wrapper", "keywords": ["Paybox", "Direct", "API", "Symfony"], "type": "library", "license": "MIT", "authors": [ { "name": "Sullivan SENECHAL", "email": "soullivaneuh@gmail.com" } ], "require": { "php": "^5.5 || ^7.0", "guzzlehttp/guzzle": "^6.0", "symfony/options-resolver": "^2.7 || ^3.0" }, "require-dev": { "phpunit/phpunit": "^5.3", "sllh/php-cs-fixer-styleci-bridge": "^2.0" },
130cbf5603488eeea3fe6394353bad93ed1f7dd7
composer.json
composer.json
JSON
<|file_sep|>original/README.md # HAProxy Documentation Converter Made to convert the HAProxy documentation into HTML. More than HTML, the main goal is to provide easy navigation. ## Documentations A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations. Converted documentations are then stored online : - [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) - [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) ## Contribute The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better. Feel free to report feature requests or to provide patches ! <|file_sep|>current/README.md # HAProxy Documentation Converter Made to convert the HAProxy documentation into HTML. More than HTML, the main goal is to provide easy navigation. ## Documentations A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations. Converted documentations are then stored online : - [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) - [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) ## Contribute The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better. Feel free to report feature requests or to provide patches ! <|file_sep|>updated/README.md
# HAProxy Documentation Converter Made to convert the HAProxy documentation into HTML. More than HTML, the main goal is to provide easy navigation. ## Documentations A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations. Converted documentations are then stored online : - HAProxy 1.4 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.4.html) - HAProxy 1.5 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.5.html) ## Contribute The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better. Feel free to report feature requests or to provide patches !
<|file_sep|>original/README.md # HAProxy Documentation Converter Made to convert the HAProxy documentation into HTML. More than HTML, the main goal is to provide easy navigation. ## Documentations A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations. Converted documentations are then stored online : - [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) - [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) ## Contribute The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better. Feel free to report feature requests or to provide patches ! <|file_sep|>current/README.md # HAProxy Documentation Converter Made to convert the HAProxy documentation into HTML. More than HTML, the main goal is to provide easy navigation. ## Documentations A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations. Converted documentations are then stored online : - [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) - [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) ## Contribute The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better. Feel free to report feature requests or to provide patches ! <|file_sep|>updated/README.md # HAProxy Documentation Converter Made to convert the HAProxy documentation into HTML. More than HTML, the main goal is to provide easy navigation. ## Documentations A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations. Converted documentations are then stored online : - HAProxy 1.4 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.4.html) - HAProxy 1.5 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.5.html) ## Contribute The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better. Feel free to report feature requests or to provide patches !
584104a2b67cb2e775987a8321349b8e64341e66
README.md
README.md
Markdown
<|file_sep|>original/zeus/utils/revisions.py ) -> Revision: """ Attempt to transform a a commit-like reference into a valid revision. """ # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision result = next(vcs_client.log(repository.id, parent=ref, limit=1)) revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision <|file_sep|>current/zeus/utils/revisions.py ) -> Revision: """ Attempt to transform a a commit-like reference into a valid revision. """ # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision result = next(vcs_client.log(repository.id, parent=ref, limit=1)) revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision <|file_sep|>updated/zeus/utils/revisions.py
# try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision try: result = vcs_client.log(repository.id, parent=ref, limit=1)[0] except IndexError: raise UnknownRevision revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision
<|file_sep|>original/zeus/utils/revisions.py ) -> Revision: """ Attempt to transform a a commit-like reference into a valid revision. """ # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision result = next(vcs_client.log(repository.id, parent=ref, limit=1)) revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision <|file_sep|>current/zeus/utils/revisions.py ) -> Revision: """ Attempt to transform a a commit-like reference into a valid revision. """ # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision result = next(vcs_client.log(repository.id, parent=ref, limit=1)) revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision <|file_sep|>updated/zeus/utils/revisions.py # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision try: result = vcs_client.log(repository.id, parent=ref, limit=1)[0] except IndexError: raise UnknownRevision revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision
209fef39f72a625e154f4455eaa6754d6a85e98b
zeus/utils/revisions.py
zeus/utils/revisions.py
Python
<|file_sep|>original/package.json { "name": "time-logger", "version": "1.0.0", "description": "", "main": "index.js", "scripts": { "test": "echo \"Error: no test specified\" && exit 1" }, "repository": { "type": "git", "url": "git+https://github.com/tabalinas/time-logger.git" }, "author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)", "license": "MIT", "bugs": { "url": "https://github.com/tabalinas/time-logger/issues" }, "homepage": "https://github.com/tabalinas/time-logger#readme" } <|file_sep|>current/package.json { "name": "time-logger", "version": "1.0.0", "description": "", "main": "index.js", "scripts": { "test": "echo \"Error: no test specified\" && exit 1" }, "repository": { "type": "git", "url": "git+https://github.com/tabalinas/time-logger.git" }, "author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)", "license": "MIT", "bugs": { "url": "https://github.com/tabalinas/time-logger/issues" }, "homepage": "https://github.com/tabalinas/time-logger#readme" } <|file_sep|>updated/package.json
"description": "", "main": "index.js", "scripts": { "test": "echo \"Error: no test specified\" && exit 1" }, "repository": { "type": "git", "url": "git+https://github.com/tabalinas/time-logger.git" }, "author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)", "license": "MIT", "bugs": { "url": "https://github.com/tabalinas/time-logger/issues" }, "homepage": "https://github.com/tabalinas/time-logger#readme", "dependencies": { "babel-standalone": "^6.21.1-0", "react": "^15.4.1", "react-dom": "^15.4.1" } }
<|file_sep|>original/package.json { "name": "time-logger", "version": "1.0.0", "description": "", "main": "index.js", "scripts": { "test": "echo \"Error: no test specified\" && exit 1" }, "repository": { "type": "git", "url": "git+https://github.com/tabalinas/time-logger.git" }, "author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)", "license": "MIT", "bugs": { "url": "https://github.com/tabalinas/time-logger/issues" }, "homepage": "https://github.com/tabalinas/time-logger#readme" } <|file_sep|>current/package.json { "name": "time-logger", "version": "1.0.0", "description": "", "main": "index.js", "scripts": { "test": "echo \"Error: no test specified\" && exit 1" }, "repository": { "type": "git", "url": "git+https://github.com/tabalinas/time-logger.git" }, "author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)", "license": "MIT", "bugs": { "url": "https://github.com/tabalinas/time-logger/issues" }, "homepage": "https://github.com/tabalinas/time-logger#readme" } <|file_sep|>updated/package.json "description": "", "main": "index.js", "scripts": { "test": "echo \"Error: no test specified\" && exit 1" }, "repository": { "type": "git", "url": "git+https://github.com/tabalinas/time-logger.git" }, "author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)", "license": "MIT", "bugs": { "url": "https://github.com/tabalinas/time-logger/issues" }, "homepage": "https://github.com/tabalinas/time-logger#readme", "dependencies": { "babel-standalone": "^6.21.1-0", "react": "^15.4.1", "react-dom": "^15.4.1" } }
3a99f1c29284f0dd05db477080a117fc9699ce84
package.json
package.json
JSON
<|file_sep|>original/Sample-App/scripts/tsconfig.json <|file_sep|>current/Sample-App/scripts/tsconfig.json <|file_sep|>updated/Sample-App/scripts/tsconfig.json
{ "compilerOptions": { "target": "es5", "out": "www/scripts/appBundle.js", "sourceMap": true, "removeComments": true, "sourceRoot": "/" } }
<|file_sep|>original/Sample-App/scripts/tsconfig.json <|file_sep|>current/Sample-App/scripts/tsconfig.json <|file_sep|>updated/Sample-App/scripts/tsconfig.json { "compilerOptions": { "target": "es5", "out": "www/scripts/appBundle.js", "sourceMap": true, "removeComments": true, "sourceRoot": "/" } }
e612feecb82179877ce5abb959e42498eb138806
Sample-App/scripts/tsconfig.json
Sample-App/scripts/tsconfig.json
JSON
<|file_sep|>docs/4.0/extend/icons.md.diff original: updated: <|file_sep|>docs/4.0/extend/icons.md.diff original: - [Entypo](http://www.entypo.com/) updated: <|file_sep|>original/docs/4.0/extend/icons.md --- Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support. ## Preferred We've tested and used these icon sets ourselves. - [Iconic](https://useiconic.com/open/) - [Octicons](https://octicons.github.com/) - [Entypo](http://www.entypo.com/) ## More options While we haven't tried these out, they do look promising and provide multiple formats—including SVG. - [Bytesize](https://github.com/danklammer/bytesize-icons) - [Google Material icons](https://material.io/icons/) - [Ionicons](http://ionicons.com/) - [Feather](https://feathericons.com/) - [Dripicons](http://demo.amitjakhu.com/dripicons/) - [Ikons](http://ikons.piotrkwiatkowski.co.uk/) - [Glyph](http://glyph.smarticons.co/) <|file_sep|>current/docs/4.0/extend/icons.md --- Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support. ## Preferred We've tested and used these icon sets ourselves. - [Iconic](https://useiconic.com/open/) - [Octicons](https://octicons.github.com/) ## More options While we haven't tried these out, they do look promising and provide multiple formats—including SVG. - [Bytesize](https://github.com/danklammer/bytesize-icons) - [Google Material icons](https://material.io/icons/) - [Ionicons](http://ionicons.com/) - [Feather](https://feathericons.com/) - [Dripicons](http://demo.amitjakhu.com/dripicons/) - [Ikons](http://ikons.piotrkwiatkowski.co.uk/) - [Glyph](http://glyph.smarticons.co/) <|file_sep|>updated/docs/4.0/extend/icons.md
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support. ## Preferred We've tested and used these icon sets ourselves. - [Iconic](https://useiconic.com/open/) - [Octicons](https://octicons.github.com/) ## More options While we haven't tried these out, they do look promising and provide multiple formats—including SVG. - [Bytesize](https://github.com/danklammer/bytesize-icons) - [Google Material icons](https://material.io/icons/) - [Ionicons](http://ionicons.com/) - [Feather](https://feathericons.com/) - [Dripicons](http://demo.amitjakhu.com/dripicons/) - [Ikons](http://ikons.piotrkwiatkowski.co.uk/) - [Glyph](http://glyph.smarticons.co/)
<|file_sep|>docs/4.0/extend/icons.md.diff original: updated: <|file_sep|>docs/4.0/extend/icons.md.diff original: - [Entypo](http://www.entypo.com/) updated: <|file_sep|>original/docs/4.0/extend/icons.md --- Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support. ## Preferred We've tested and used these icon sets ourselves. - [Iconic](https://useiconic.com/open/) - [Octicons](https://octicons.github.com/) - [Entypo](http://www.entypo.com/) ## More options While we haven't tried these out, they do look promising and provide multiple formats—including SVG. - [Bytesize](https://github.com/danklammer/bytesize-icons) - [Google Material icons](https://material.io/icons/) - [Ionicons](http://ionicons.com/) - [Feather](https://feathericons.com/) - [Dripicons](http://demo.amitjakhu.com/dripicons/) - [Ikons](http://ikons.piotrkwiatkowski.co.uk/) - [Glyph](http://glyph.smarticons.co/) <|file_sep|>current/docs/4.0/extend/icons.md --- Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support. ## Preferred We've tested and used these icon sets ourselves. - [Iconic](https://useiconic.com/open/) - [Octicons](https://octicons.github.com/) ## More options While we haven't tried these out, they do look promising and provide multiple formats—including SVG. - [Bytesize](https://github.com/danklammer/bytesize-icons) - [Google Material icons](https://material.io/icons/) - [Ionicons](http://ionicons.com/) - [Feather](https://feathericons.com/) - [Dripicons](http://demo.amitjakhu.com/dripicons/) - [Ikons](http://ikons.piotrkwiatkowski.co.uk/) - [Glyph](http://glyph.smarticons.co/) <|file_sep|>updated/docs/4.0/extend/icons.md Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support. ## Preferred We've tested and used these icon sets ourselves. - [Iconic](https://useiconic.com/open/) - [Octicons](https://octicons.github.com/) ## More options While we haven't tried these out, they do look promising and provide multiple formats—including SVG. - [Bytesize](https://github.com/danklammer/bytesize-icons) - [Google Material icons](https://material.io/icons/) - [Ionicons](http://ionicons.com/) - [Feather](https://feathericons.com/) - [Dripicons](http://demo.amitjakhu.com/dripicons/) - [Ikons](http://ikons.piotrkwiatkowski.co.uk/) - [Glyph](http://glyph.smarticons.co/)
dd041e7bd5760c74b23a9ca1e9a76dc3de1d7e24
docs/4.0/extend/icons.md
docs/4.0/extend/icons.md
Markdown
<|file_sep|>original/src/CMakeLists.txt # # Building libreset # # # We use the var SOURCE_FILES to hold all source files to be linked together # into libreset. Additionally, checks checky may be performed on that source # files. # set(SOURCE_FILES libreset/ll.c ) # # Libreset will be a shared object # add_library(reset SHARED ${SOURCE_FILES}) <|file_sep|>current/src/CMakeLists.txt # # Building libreset # # # We use the var SOURCE_FILES to hold all source files to be linked together # into libreset. Additionally, checks checky may be performed on that source # files. # set(SOURCE_FILES libreset/ll.c ) # # Libreset will be a shared object # add_library(reset SHARED ${SOURCE_FILES}) <|file_sep|>updated/src/CMakeLists.txt
# # We use the var SOURCE_FILES to hold all source files to be linked together # into libreset. Additionally, checks checky may be performed on that source # files. # set(SOURCE_FILES libreset/ll.c ) # # General compile options # add_compile_options(-Wall -Wextra -Wunused -Wformat -Wno-unused-parameter -pedantic) # # Libreset will be a shared object # add_library(reset SHARED ${SOURCE_FILES})
<|file_sep|>original/src/CMakeLists.txt # # Building libreset # # # We use the var SOURCE_FILES to hold all source files to be linked together # into libreset. Additionally, checks checky may be performed on that source # files. # set(SOURCE_FILES libreset/ll.c ) # # Libreset will be a shared object # add_library(reset SHARED ${SOURCE_FILES}) <|file_sep|>current/src/CMakeLists.txt # # Building libreset # # # We use the var SOURCE_FILES to hold all source files to be linked together # into libreset. Additionally, checks checky may be performed on that source # files. # set(SOURCE_FILES libreset/ll.c ) # # Libreset will be a shared object # add_library(reset SHARED ${SOURCE_FILES}) <|file_sep|>updated/src/CMakeLists.txt # # We use the var SOURCE_FILES to hold all source files to be linked together # into libreset. Additionally, checks checky may be performed on that source # files. # set(SOURCE_FILES libreset/ll.c ) # # General compile options # add_compile_options(-Wall -Wextra -Wunused -Wformat -Wno-unused-parameter -pedantic) # # Libreset will be a shared object # add_library(reset SHARED ${SOURCE_FILES})
2ddd0aa55f44a25dfe08bc3dc6b01e9d8514520c
src/CMakeLists.txt
src/CMakeLists.txt
Text
<|file_sep|>original/coverage.bash #!/bin/bash base_package=github.com/mitsuse/bullet base_path=${GOPATH}/src/${base_package} package_list=( ${base_package} ) if [ ! -d ${base_path}/coverprofile ] then mkdir ${base_path}/coverprofile else rm ${base_path}/coverprofile/*.coverprofile fi for package in ${package_list[@]} do cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile cover_path=${base_path}/coverprofile/${cover_name} go test -covermode=count -coverprofile ${cover_path} ${package} <|file_sep|>current/coverage.bash #!/bin/bash base_package=github.com/mitsuse/bullet base_path=${GOPATH}/src/${base_package} package_list=( ${base_package} ) if [ ! -d ${base_path}/coverprofile ] then mkdir ${base_path}/coverprofile else rm ${base_path}/coverprofile/*.coverprofile fi for package in ${package_list[@]} do cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile cover_path=${base_path}/coverprofile/${cover_name} go test -covermode=count -coverprofile ${cover_path} ${package} <|file_sep|>updated/coverage.bash
#!/bin/bash base_package=github.com/mitsuse/bullet base_path=${GOPATH}/src/${base_package} package_list=( ${base_package} ${base_package}/pushbullet ${base_package}/pushbullet/pushes ) if [ ! -d ${base_path}/coverprofile ] then mkdir ${base_path}/coverprofile else rm ${base_path}/coverprofile/*.coverprofile fi for package in ${package_list[@]} do cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile
<|file_sep|>original/coverage.bash #!/bin/bash base_package=github.com/mitsuse/bullet base_path=${GOPATH}/src/${base_package} package_list=( ${base_package} ) if [ ! -d ${base_path}/coverprofile ] then mkdir ${base_path}/coverprofile else rm ${base_path}/coverprofile/*.coverprofile fi for package in ${package_list[@]} do cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile cover_path=${base_path}/coverprofile/${cover_name} go test -covermode=count -coverprofile ${cover_path} ${package} <|file_sep|>current/coverage.bash #!/bin/bash base_package=github.com/mitsuse/bullet base_path=${GOPATH}/src/${base_package} package_list=( ${base_package} ) if [ ! -d ${base_path}/coverprofile ] then mkdir ${base_path}/coverprofile else rm ${base_path}/coverprofile/*.coverprofile fi for package in ${package_list[@]} do cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile cover_path=${base_path}/coverprofile/${cover_name} go test -covermode=count -coverprofile ${cover_path} ${package} <|file_sep|>updated/coverage.bash #!/bin/bash base_package=github.com/mitsuse/bullet base_path=${GOPATH}/src/${base_package} package_list=( ${base_package} ${base_package}/pushbullet ${base_package}/pushbullet/pushes ) if [ ! -d ${base_path}/coverprofile ] then mkdir ${base_path}/coverprofile else rm ${base_path}/coverprofile/*.coverprofile fi for package in ${package_list[@]} do cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile
8ae118a96e9c5b009e92bb694dc914108636edf5
coverage.bash
coverage.bash
Shell
<|file_sep|>resources/views/admin/modalities_index.blade.php.diff original: updated: <p>Click a modality to edit</p> <|file_sep|>resources/views/admin/modalities_index.blade.php.diff original: <td>{{ $modality->modality }}</td> <form action="/admin/modalities/{{ $modality->id }}/edit" method="POST"> <button type="submit">Edit</button> </form> updated: <td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td> <|file_sep|>original/resources/views/admin/modalities_index.blade.php <tbody> @foreach ($modalities->chunk(2) as $chunk ) <tr> @foreach ($chunk as $modality) <td>{{ $modality->id }}</td> <td>{{ $modality->modality }}</td> <form action="/admin/modalities/{{ $modality->id }}/edit" method="POST"> <button type="submit">Edit</button> </form> @endforeach </tr> @endforeach </tbody> </table> <h2>Add a Modality</h2> <!-- Add a new modality --> <form action="/admin/modalities" method="POST"> {{ csrf_field() }} New Modality: <input type="TEXT" name="modality" size="25" /> <button type="SUBMIT">Add modality</button> / <a href="/">Main</a> <|file_sep|>current/resources/views/admin/modalities_index.blade.php <tbody> @foreach ($modalities->chunk(2) as $chunk ) <tr> @foreach ($chunk as $modality) <td>{{ $modality->id }}</td> <td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td> @endforeach </tr> @endforeach </tbody> </table> <h2>Add a Modality</h2> <!-- Add a new modality --> <form action="/admin/modalities" method="POST"> {{ csrf_field() }} New Modality: <input type="TEXT" name="modality" size="25" /> <button type="SUBMIT">Add modality</button> / <a href="/">Main</a> </form> @endsection <|file_sep|>updated/resources/views/admin/modalities_index.blade.php
<tr> @foreach ($chunk as $modality) <td>{{ $modality->id }}</td> <td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td> @endforeach </tr> @endforeach </tbody> </table> <h2>Add a Modality</h2> <!-- Add a new modality --> <form class="form-inline" action="/admin/modalities" method="POST"> <div class="form-group"> {{ csrf_field() }} <label for="modality">New Modality:</label> <input type="TEXT" class="form-control" id="modality" name="modality" size="25" /> <button class="btn btn-default" type="SUBMIT">Add modality</button> / <a href="/">Main</a> </div> </form> @endsection
<|file_sep|>resources/views/admin/modalities_index.blade.php.diff original: updated: <p>Click a modality to edit</p> <|file_sep|>resources/views/admin/modalities_index.blade.php.diff original: <td>{{ $modality->modality }}</td> <form action="/admin/modalities/{{ $modality->id }}/edit" method="POST"> <button type="submit">Edit</button> </form> updated: <td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td> <|file_sep|>original/resources/views/admin/modalities_index.blade.php <tbody> @foreach ($modalities->chunk(2) as $chunk ) <tr> @foreach ($chunk as $modality) <td>{{ $modality->id }}</td> <td>{{ $modality->modality }}</td> <form action="/admin/modalities/{{ $modality->id }}/edit" method="POST"> <button type="submit">Edit</button> </form> @endforeach </tr> @endforeach </tbody> </table> <h2>Add a Modality</h2> <!-- Add a new modality --> <form action="/admin/modalities" method="POST"> {{ csrf_field() }} New Modality: <input type="TEXT" name="modality" size="25" /> <button type="SUBMIT">Add modality</button> / <a href="/">Main</a> <|file_sep|>current/resources/views/admin/modalities_index.blade.php <tbody> @foreach ($modalities->chunk(2) as $chunk ) <tr> @foreach ($chunk as $modality) <td>{{ $modality->id }}</td> <td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td> @endforeach </tr> @endforeach </tbody> </table> <h2>Add a Modality</h2> <!-- Add a new modality --> <form action="/admin/modalities" method="POST"> {{ csrf_field() }} New Modality: <input type="TEXT" name="modality" size="25" /> <button type="SUBMIT">Add modality</button> / <a href="/">Main</a> </form> @endsection <|file_sep|>updated/resources/views/admin/modalities_index.blade.php <tr> @foreach ($chunk as $modality) <td>{{ $modality->id }}</td> <td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td> @endforeach </tr> @endforeach </tbody> </table> <h2>Add a Modality</h2> <!-- Add a new modality --> <form class="form-inline" action="/admin/modalities" method="POST"> <div class="form-group"> {{ csrf_field() }} <label for="modality">New Modality:</label> <input type="TEXT" class="form-control" id="modality" name="modality" size="25" /> <button class="btn btn-default" type="SUBMIT">Add modality</button> / <a href="/">Main</a> </div> </form> @endsection
664be3667730e7e723fb4d50844da530640e66a3
resources/views/admin/modalities_index.blade.php
resources/views/admin/modalities_index.blade.php
PHP