patch stringlengths 18 160k | callgraph stringlengths 4 179k | summary stringlengths 4 947 | msg stringlengths 6 3.42k |
|---|---|---|---|
@@ -483,6 +483,16 @@ module Engine
logo: '1835/PR',
simple_logo: '1835/PR.alt',
tokens: [0],
+ abilities: [
+ {
+ type: 'exchange',
+ corporations: %w[PR_12],
+ owner_type: 'player',
+ when: ['Phase 2.3', 'Phase 2.4', 'Phase 3.1'],
+ # reserved papers perhaps a option
+ from: 'ipo',
+ },
+ ],
coordinates: 'C11',
color: '#37383a',
},
| [Game->[operating_round->[new],include_meta,register_colors,freeze],require_relative] | Return a list of colors for a specific residue. A list of all possible values of a variable in a standardised format. Devuelve un objeto nueva nueva nueva nueva. | I'm not entirely sure this when is right, but can be fixed later. |
@@ -381,9 +381,9 @@ int ssl3_get_record(SSL *s)
goto f_err;
}
#ifdef SSL_DEBUG
- printf("dec %d\n", rr->length);
+ printf("dec %ld\n", rr->length);
{
- unsigned int z;
+ size_t z;
for (z = 0; z < rr->length; z++)
printf("%02X%c", rr->data[z], ((z + 1) % 16) ? ' ' : '\n');
}
| [No CFG could be retrieved] | Decode and decrypt a record. RFC Section 5. 2. 5. 2. 5. 1. | %ld is not the right specifier for a size_t. %d is a signed type while size_t is unsigned, and a size_t doesn't have to have the same size as a long and doesn't at least on 64 bit windows. In C99 you should use %zu. We probably need something like PRIu32 for platforms that don't support %z. |
@@ -599,6 +599,14 @@ func convertWindowsProfileToV20170701(api *WindowsProfile, v20170701Profile *v20
func convertWindowsProfileToVLabs(api *WindowsProfile, vlabsProfile *vlabs.WindowsProfile) {
vlabsProfile.AdminUsername = api.AdminUsername
vlabsProfile.AdminPassword = api.AdminPassword
+ if api.ImageRef != nil {
+ vlabsProfile.ImageRef = &vlabs.ImageReference{}
+ vlabsProfile.ImageRef.Gallery = api.ImageRef.Gallery
+ vlabsProfile.ImageRef.Name = api.ImageRef.Name
+ vlabsProfile.ImageRef.ResourceGroup = api.ImageRef.ResourceGroup
+ vlabsProfile.ImageRef.SubscriptionID = api.ImageRef.SubscriptionID
+ vlabsProfile.ImageRef.Version = api.ImageRef.Version
+ }
vlabsProfile.ImageVersion = api.ImageVersion
vlabsProfile.WindowsImageSourceURL = api.WindowsImageSourceURL
vlabsProfile.WindowsPublisher = api.WindowsPublisher
| [SetSubnetIPv6,Distro,DependenciesLocation,KubeProxyMode,MatchString,Sprintf,SetSubnet,OSType,AgentPoolProfileRole,ProvisioningState,Make,HasPrefix,IsDCOS] | convertWindowsProfileToV20160930 - convert WindowsProfile to WindowsProfile. convertOrchestratorProfileToV20160930 - converts v20160930. | if this is showing up as negative coverage, it might need one of the test cases that compares the final arm template against a known string |
@@ -90,7 +90,15 @@ class WebsiteSearchController
$hits = $this->searchManager
->createSearch($queryString)
->locale($locale)
- ->index('page_' . $webspace->getKey() . '_published')
+ ->indexes(
+ \str_replace(
+ '#webspace#',
+ $webspace->getKey(),
+ \array_filter(
+ \array_values($this->indexes)
+ )
+ )
+ )
->execute();
$template = $webspace->getTemplate('search', $request->getRequestFormat());
| [WebsiteSearchController->[queryAction->[resolve,render,getMimeType,getTemplate,getLocale,getWebspace,set,exists,getRequestParameter,getRequestFormat,execute]]] | Queries a node in the node list. | I personally would move this logic into the `SuluSearchExtension.php`, where the parameter is set. |
@@ -251,6 +251,10 @@ func initializeORM(config *orm.Config, shutdownSignal gracefulpanic.Signal) (*or
if err != nil {
return nil, errors.Wrap(err, "initializeORM#NewORM")
}
+ if config.DatabaseBackupEnabled() {
+ databaseBackup := periodicbackup.NewDatabaseBackup(config.DatabaseBackupFrequency(), config.DatabaseURL(), config.RootDir(), logger.Default)
+ databaseBackup.RunBackupGracefully()
+ }
if err = CheckSquashUpgrade(orm.DB); err != nil {
panic(err)
}
| [Unscoped->[Unscoped],Close->[Close],AuthorizedUserWithSession->[AuthorizedUserWithSession],ImportKey->[SyncDiskKeyStoreToDB]] | initializeORM initializes the ORM with the necessary information. | not sure if it's the best place to run the backup, but it's just before migrations |
@@ -120,8 +120,11 @@ class Tester(unittest.TestCase):
res = Image.fromarray(result.permute(1, 2, 0).contiguous().numpy())
res.save(path)
- expected = torch.as_tensor(np.array(Image.open(path))).permute(2, 0, 1)
- self.assertTrue(torch.equal(result, expected))
+ if PILLOW_VERSION >= (8, 2):
+ # The reference image is only valid for new PIL versions
+ expected = torch.as_tensor(np.array(Image.open(path))).permute(2, 0, 1)
+ self.assertTrue(torch.equal(result, expected))
+
# Check if modification is not in place
self.assertTrue(torch.all(torch.eq(boxes, boxes_cp)).item())
self.assertTrue(torch.all(torch.eq(img, img_cp)).item())
| [Tester->[test_draw_invalid_masks->[full,assertRaises],test_save_image->[assertTrue,rand,NamedTemporaryFile,save_image,exists],test_normalize_in_make_grid->[assertTrue,rand,round,tensor,max,min,make_grid,equal],test_draw_boxes->[full,fromarray,abspath,assertTrue,all,dirname,clone,draw_bounding_boxes,as_tensor,permute,join,exists,save,equal],test_draw_segmentation_masks_no_colors->[full,fromarray,abspath,assertTrue,all,dirname,clone,as_tensor,permute,join,exists,save,draw_segmentation_masks,equal],test_draw_segmentation_masks_colors->[full,fromarray,abspath,assertTrue,all,dirname,clone,as_tensor,permute,join,exists,save,draw_segmentation_masks,equal],test_save_image_single_pixel_file_object->[assertTrue,rand,NamedTemporaryFile,save_image,BytesIO,to_tensor,open,equal],test_draw_boxes_vanilla->[full,fromarray,abspath,assertTrue,all,dirname,clone,draw_bounding_boxes,as_tensor,permute,join,exists,save,equal],test_draw_invalid_boxes->[full,tensor,assertRaises],test_save_image_single_pixel->[assertTrue,rand,NamedTemporaryFile,save_image,exists],test_save_image_file_object->[assertTrue,rand,NamedTemporaryFile,save_image,BytesIO,to_tensor,open,equal],test_make_grid_not_inplace->[assertTrue,rand,clone,make_grid,equal],skipIf],tensor,main] | Test drawing of boxes. Check if result is not in place or not. | Another solution is to set the condition to `if PILLOW_VERSION < (8, 2)` and not update the image. |
@@ -0,0 +1,15 @@
+import javax.servlet.http.Cookie;
+import java.net.HttpCookie;
+class A {
+ void foo() {
+ Cookie myCookie = new Cookie("name", "val");
+ myCookie.setDomain(".com"); // Noncompliant [[sc=23;ec=29]] {{Specify at least a second-level cookie domain.}}
+ myCookie.setDomain(".myDomain.com"); // Compliant
+ }
+ void HttpCookie(String domain) {
+ HttpCookie myCookie = new Cookie("name", "val");
+ myCookie.setDomain(".com"); // Noncompliant
+ myCookie.setDomain(".myDomain.com"); // Compliant
+ myCookie.setDomain(domain); // Compliant
+ }
+}
| [No CFG could be retrieved] | No Summary Found. | Should it not be `new HttpCookie` instead? |
@@ -375,6 +375,13 @@ static WRITE_TRAN ossl_statem_client13_write_transition(SSL *s)
* ossl_statem_client_write_transition().
*/
switch (st->hand_state) {
+ case TLS_ST_CR_CERT_REQ:
+ if (s->post_handshake_auth == SSL_PHA_REQUESTED) {
+ st->hand_state = TLS_ST_CW_CERT;
+ return WRITE_TRAN_CONTINUE;
+ }
+ /* Fall through */
+
default:
/* Shouldn't happen */
SSLfatal(s, SSL_AD_INTERNAL_ERROR,
| [No CFG could be retrieved] | The function for the client - side write - transition. The handshake state is one of TLS_ST_CW_CHANGE TLS_ST_CW. | side note: I'm a little reluctant to do a fallthrough-to-default here, since though it does make for shorter error handling in this case, we can only pull this trick once, and there's not anything particularly special about CERT_REQ that should give it this privileged status. |
@@ -231,6 +231,18 @@ namespace ILCompiler
try
{
EcmaModule module = typeSystemContext.GetModuleFromPath(referenceFile);
+ if (versionBubbleModulesHash.Contains(module))
+ {
+ // Ignore reference assemblies that have also been passed as inputs
+ continue;
+ }
+ if (_commandLineOptions.Composite)
+ {
+ // In composite mode, add reference assemblies as inputs but don't root all their methods
+ // This will need more granularity (probably two types of references) when implementing
+ // shared framework / multi-layer composite R2R files.
+ inputModules.Add(module);
+ }
referenceableModules.Add(module);
if (_commandLineOptions.InputBubble)
{
| [Program->[Run->[ProcessCommandLine,InitializeDefaultOptions],InnerMain->[Run,DumpReproArguments]]] | Runs the type system context. Checks if a type can be found in the input file and if so determines if a type Adds a new module group to the list of modules that can be compiled. Compile a single object if it exists. | It seems very strange that we'd want to include in the input set any modules that were only specified as references. I don't see why you are doing this. Instead I would expect to see the logic that changed what were the reference vs input modules to be placed into SuperILC |
@@ -52,12 +52,9 @@ def AssembleTestSuites():
smallSuite = suites['small']
smallSuite.addTest(CouetteFlowTest('testCouetteFlow2DSymbolicStokes'))
smallSuite.addTest(CouetteFlowTest('testCouetteFlow2DWeaklyCompressibleNavierStokes'))
- smallSuite.addTest(EmbeddedCouetteTest('testEmbeddedCouette2D'))
- smallSuite.addTest(EmbeddedCouetteTest('testEmbeddedSlipCouette2D'))
- smallSuite.addTest(EmbeddedCouetteTest('testEmbeddedAusasCouette2D'))
- smallSuite.addTest(EmbeddedCouetteTest('testEmbeddedDevelopmentCouette2D'))
- smallSuite.addTest(EmbeddedCouetteTest('testEmbeddedSlipDevelopmentCouette2D'))
- smallSuite.addTest(EmbeddedCouetteTest('testEmbeddedAusasDevelopmentCouette2D'))
+ smallSuite.addTest(EmbeddedCouetteFlowTest('testEmbeddedCouetteFlowEmbeddedWeaklyCompressibleNavierStokes2D'))
+ smallSuite.addTest(EmbeddedCouetteFlowTest('testEmbeddedCouetteFlowEmbeddedWeaklyCompressibleNavierStokes2DSlip'))
+ smallSuite.addTest(EmbeddedCouetteFlowTest('testEmbeddedCouetteFlowEmbeddedWeaklyCompressibleNavierStokesDiscontinuous2D'))
smallSuite.addTest(EmbeddedCouetteImposedTest('testEmbeddedCouetteImposed2D'))
smallSuite.addTest(EmbeddedPistonTest('testEmbeddedPiston2D'))
smallSuite.addTest(EmbeddedReservoirTest('testEmbeddedReservoir2D'))
| [AssembleTestSuites->[BuoyancyTest,CouetteFlowTest,DarcyChannelTest,FluidAnalysisTest,FluidElementTest,EmbeddedVelocityInletEmulationTest,EmbeddedCouetteImposedTest,AdjointFluidTest,ManufacturedSolutionTest,AdjointVMSSensitivity2D,ConsistentLevelsetNodalGradientTest,NavierStokesWallConditionTest,HDF5IOTest,EmbeddedPistonTest,EmbeddedCouetteTest,TestLoader,EmbeddedReservoirTest,SodShockTubeTest,addTest,ArtificialCompressibilityTest,EmbeddedReservoirDiscontinuousTest,AdjointVMSElement2D,addTests],PrintInfo,SetVerbosity,AssembleTestSuites,GetDefaultOutput,runTests,RunTestSuite] | AssembleTestSuites - Creates a test suite with the selected tests. Adds tests to the Nightly test suite with the selected tests. Adds tests to the NightSuite object. Load all tests that are not in nighly and return a test suite that contains all. | These are old tests which are removed ryt? |
@@ -22,4 +22,11 @@ public interface WithVersion {
default int getVersion() {
return 1;
}
+
+ /**
+ * Marker interface to indicate that the revision should be automatically updated
+ * by the data manager
+ */
+ interface AutoUpdatable {
+ }
}
| [No CFG could be retrieved] | Returns the version of the . | Naming things is hard, this is a good name, might `Current` be a better one? So `WithVersion.Current`? |
@@ -174,7 +174,7 @@ class RepoUnitAssociationManager(object):
manager_factory.repo_manager().update_unit_count(
repo_id, unique_count)
- def associate_from_repo(self, source_repo_id, dest_repo_id, criteria=None):
+ def associate_from_repo(self, source_repo_id, dest_repo_id, criteria=None, import_config_override=None):
"""
Creates associations in a repository based on the contents of a source
repository. Units from the source repository can be filtered by
| [remove_from_importer->[create_transfer_units,calculate_associated_type_ids],RepoUnitAssociationManager->[associate_all_by_ids->[associate_unit_by_id]]] | Associates all units in a given list with the given type and content units. This method is called by the base repository to associate the given node node with the destination node Imports all of the units from source repository to destination repository. | The name is a bit misleading. "copy_config_override" might have been more appropriate since it's a copy configuration. Or "importer_config_override". But ultimately not a big deal and not a blocker for this pull request. |
@@ -232,6 +232,11 @@ public final class HiveSessionProperties
"Parquet: Fail when scanning Parquet files with corrupted statistics",
hiveClientConfig.isFailOnCorruptedParquetStatistics(),
false),
+ dataSizeSessionProperty(
+ PARQUET_MAX_READ_BLOCK_SIZE,
+ "Parquet: Maximum size of a block to read",
+ hiveClientConfig.getParquetMaxReadBlockSize(),
+ false),
dataSizeSessionProperty(
PARQUET_WRITER_BLOCK_SIZE,
"Parquet: Writer block size",
| [HiveSessionProperties->[getOrcOptimizedWriterValidateMode->[valueOf],getHiveStorageFormat->[valueOf],dataSizeSessionProperty->[valueOf],valueOf->[valueOf],valueOf]] | Experimental - related configuration options. Experimental - only function. | Is this a true "max" or a "soft max" as the ORC reader has? |
@@ -232,7 +232,13 @@ class GroupedShuffleRangeTracker(iobase.RangeTracker):
else 'non-')))
def try_claim(self, decoded_group_start):
+ assert not self.done()
with self._lock:
+ if self.done():
+ raise ValueError('Trying to claim record at %r after already closing '
+ 'RangeTracker by invoking \'set_done()\'',
+ decoded_group_start)
+
self._validate_decoded_group_start(decoded_group_start, True)
if (self.stop_position()
and decoded_group_start >= self.stop_position()):
| [UnsplittableRangeTracker->[position_at_fraction->[position_at_fraction],start_position->[start_position],try_claim->[try_claim],fraction_consumed->[fraction_consumed],stop_position->[stop_position],set_current_position->[set_current_position]],GroupedShuffleRangeTracker->[try_split->[stop_position,start_position,last_group_start],set_current_position->[_validate_decoded_group_start],try_claim->[stop_position,_validate_decoded_group_start],_validate_decoded_group_start->[start_position,last_group_start]],OffsetRangeTracker->[position_at_fraction->[stop_position,start_position],try_claim->[stop_position,_validate_record_start],fraction_consumed->[stop_position,start_position],try_split->[stop_position,start_position],set_current_position->[_validate_record_start]]] | Try to claim a group start. | Here (and elsewhere). This assert is redundant with the check below. |
@@ -0,0 +1,18 @@
+# frozen_string_literal: true
+
+require 'snabberb/component'
+
+require 'view/part/city'
+
+module View
+ module Part
+ class Cities < Snabberb::Component
+ needs :tile
+ needs :region_use
+
+ def render
+ h(Part::City, region_use: @region_use, city: @tile.cities.first) if @tile.cities.count == 1
+ end
+ end
+ end
+end
| [No CFG could be retrieved] | No Summary Found. | should this be a Base instead of component, then you don't need tile and region_use |
@@ -684,7 +684,9 @@ export class AmpLightboxViewer extends AMP.BaseElement {
this.element.ownerDocument.body.appendChild(transLayer);
const rect = layoutRectFromDomRect(sourceImage
./*OK*/getBoundingClientRect());
- const imageBox = this.getCurrentElement_().imageViewer.getImageBox();
+
+ const imageBox = /**@type {?}*/ (this.getCurrentElement_().imageViewer)
+ .implementation_.getImageBox();
const clone = sourceImage.cloneNode(true);
clone.className = '';
| [No CFG could be retrieved] | Animation to transition in a lightboxable image. Adds a motion animation to the image. | Since the image viewer is a full custom element now, you should be able to access `getImageBox` directly if you have a reference to its DOM element. |
@@ -34,6 +34,7 @@ class Singularity(AutotoolsPackage):
version('2.4', 'd357ce68ef2f8149edd84155731531465dbe74148c37719f87f168fc39384377')
version('2.3.1', '292ff7fe3db09c854b8accf42f763f62')
+ version('master', git='https://github.com/singularityware/singularity.git')
depends_on('m4', type='build')
depends_on('autoconf', type='build')
| [Singularity->[depends_on,version]] | This site shows the version of a single - node package. | please call it `develop` to be consistent with the rest of Spack |
@@ -103,7 +103,7 @@ public class SetRequest {
return this;
}
- public Builder setFieldValues(List<Map<String, String>> fieldValues) {
+ public Builder setFieldValues(List<String> fieldValues) {
this.fieldValues = fieldValues;
return this;
}
| [SetRequest->[Builder->[build->[checkArgument,SetRequest,isEmpty],emptyList],builder->[Builder]]] | Sets the values and fieldValues of the . | I'm wondering if we should not provide instead a `setFieldValues(List<Map<String,String>> fieldValues)` to hide the JSON... But in fact, I think the goal is to map as close as possible what is done when using HTTP so I think you can forget. |
@@ -571,6 +571,13 @@ class WPSEO_Metabox extends WPSEO_Meta {
switch ( $meta_field_def['type'] ) {
case 'pageanalysis':
+ $content_analysis_active = $this->options['content_analysis_active'];
+ $keyword_analysis_active = $this->options['keyword_analysis_active'];
+
+ if( $content_analysis_active === false && $keyword_analysis_active === false ) {
+ break;
+ }
+
$content .= '<div id="pageanalysis">';
$content .= '<section class="yoast-section" id="wpseo-pageanalysis-section">';
$content .= '<h3 class="yoast-section__heading yoast-section__heading-icon yoast-section__heading-icon-list">' . __( 'Analysis', 'wordpress-seo' ) . '</h3>';
| [WPSEO_Metabox->[seo_score_posts_where->[seo_score_posts_where],add_meta_box->[is_metabox_hidden],add_custom_box->[add_meta_box],column_sort_orderby->[column_sort_orderby],get_replace_vars->[get_metabox_post],posts_filter_dropdown->[posts_filter_dropdown],column_content->[column_content],column_heading->[column_heading],publish_box->[is_metabox_hidden],column_sort->[column_sort],enqueue->[localize_replace_vars_script,localize_post_scraper_script,localize_shortcode_plugin_script,is_metabox_hidden],column_hidden->[column_hidden]]] | This function is used to generate a Yoast metabox Renders the Yoast meta field. Renders the meta field content. This function renders the Yoast meta field This function renders the Yoast checkboxes and radio buttons in the form. | Can you add a space after the `if`: `if ( $content_analysis_active === false && $keyword_analysis_active === false ) {` |
@@ -139,7 +139,6 @@ public final class HibernateOrmProcessor {
private static final DotName STATIC_METAMODEL = DotName.createSimple(StaticMetamodel.class.getName());
private static final String INTEGRATOR_SERVICE_FILE = "META-INF/services/org.hibernate.integrator.spi.Integrator";
- private static final String SERVICE_CONTRIBUTOR_SERVICE_FILE = "META-INF/services/org.hibernate.service.spi.ServiceContributor";
/**
* Hibernate ORM configuration
| [HibernateOrmProcessor->[toArray->[toArray],test->[toArray],registerBeans->[getMultiTenancyStrategy,build],createMetricBuildItem->[build],handleHibernateORMWithNoPersistenceXml->[getSqlLoadScript]]] | Creates a Hibernate ORM processor which processes the given . The Hibernate hot deployment build item. | Somehow, it looks like I missed something when I worked on that. We can probably remove it for now but at some point we might have to finish that properly. |
@@ -37,5 +37,7 @@ class PyBeautifulsoup4(Package):
extends('python')
+ depends_on('py-setuptools', type='build')
+
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix={0}'.format(prefix))
| [PyBeautifulsoup4->[install->[python,format],version,extends]] | Installs a new package. | this should probably be a `type='build'` dependency |
@@ -134,11 +134,9 @@ public class RulesAttachmentExporter extends DefaultAttachmentExporter {
try {
oValue = field.get(attachment);
} catch (final IllegalArgumentException e) {
- // TODO Auto-generated catch block
- e.printStackTrace();
+ ClientLogger.logError(e);
} catch (final IllegalAccessException e) {
- // TODO Auto-generated catch block
- e.printStackTrace();
+ ClientLogger.logError(e);
}
String intList = "";
final HashMap<Integer, Integer> intMap = (HashMap<Integer, Integer>) oValue;
| [RulesAttachmentExporter->[printOption->[printOption]]] | Method mTurnsHandler. | Please create an error message string similar to: `String failMsg = "Failed to get attachment: "+ attachment + ", from field: " + field;` Then call client logger with the error mssage `logError(failMsg, e)` Otherwise if this fails we'll be searching for how to reproduce the problem. The extra information skips debugging completely. |
@@ -14,8 +14,6 @@ import java.util.List;
public class RouteScripted extends Route {
private static final long serialVersionUID = 604474811874966546L;
- public RouteScripted() {}
-
/**
* Shameless cheating. Making a fake route, so as to handle battles properly without breaking battleTracker protected
* status or
| [RouteScripted->[getEnd->[getEnd],hasExactlyOneStep->[numberOfSteps],getTerritoryAtStep->[getEnd,getTerritoryAtStep],add->[add],getSteps->[getSteps,numberOfSteps],getMovementCost->[getMovementCost],numberOfSteps->[numberOfSteps]]] | A scripted route that can be used to route a route scripted route. Gets the territory at the given step. | I mixed scopes here a bit, the two constructors removed are unused. Side note: would be great to see this class combined with `Route.java` |
@@ -1,6 +1,8 @@
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
+__version__ = '999.0.0-developing'
+
from .env_vars import dispatcher_env_vars
if dispatcher_env_vars.SDK_PROCESS != 'dispatcher':
| [NoMoreTrialError->[__init__->[super]]] | Creates an object that represents the last non - trial error in the network. | where this string is replaced with correct version? this string in `setup.py` is replaced in `Makefile` |
@@ -148,7 +148,7 @@ func (e *Identify2WithUID) runReturnError(ctx *Context) (err error) {
return err
}
- if !e.useAnyAssertions() && e.checkFastCacheHit() && e.allowEarlyOuts() {
+ if !e.useAnyAssertions() && e.allowEarlyOuts() && e.checkFastCacheHit() {
e.G().Log.Debug("| hit fast cache")
return nil
}
| [runIdentifyUI->[maybeCacheResult,insertTrackToken,checkRemoteAssertions],toUserPlusKeys->[getNow],CCLCheckCompleted->[unblock],getTrackType->[Result],loadUsers->[loadThem,loadMe],createIdentifyState->[getTrackChainLink]] | runReturnError runs the Identify2WithUID and returns an error if there is a problem check - remote assertions - tracking check - remote assertions check - remote assertions check - remote assertions. | why this reorder? |
@@ -90,6 +90,15 @@ SYSCTL_ULONG(_vfs_zfs, OID_AUTO, abd_chunk_size, CTLFLAG_RDTUN,
kmem_cache_t *abd_chunk_cache;
static kstat_t *abd_ksp;
+/*
+ * We use a scattered SPA_MAXBLOCKSIZE sized ABD whose chunks are
+ * just a single zero'd sized zfs_abd_chunk_size buffer. This
+ * allows us to conserve memory by only using a single zero buffer
+ * for the scatter chunks.
+ */
+abd_t *abd_zero_scatter = NULL;
+static char *abd_zero_buf = NULL;
+
static void
abd_free_chunk(void *c)
{
| [No CFG could be retrieved] | - Number of blocks which are currently allocated excluding the linear ABDs which don t own zfs_abd_chunk_size - offset in bytes of zfs_ab. | I don't think this macro aids in understanding the code (and it doesn't make it any more concise either). Looks like it's barely used, so let's remove it. |
@@ -27,10 +27,11 @@
break;
default:
$new_style = 'category-products';
- }
- if (zen_get_product_types_to_category($box_categories_array[$i]['path']) == 3 or ($box_categories_array[$i]['top'] != 'true' and SHOW_CATEGORIES_SUBCATEGORIES_ALWAYS != 1)) {
- // skip if this is for the document box (==3)
- } else {
+ }
+ if ((CATEGORIES_PRODUCTS_INACTIVE_HIDE == 1 && $box_categories_array[$i]['count'] == 0) || zen_get_product_types_to_category($box_categories_array[$i]['path']) == 3 or ($box_categories_array[$i]['top'] != 'true' and SHOW_CATEGORIES_SUBCATEGORIES_ALWAYS != 1)) {
+ // skip if this is for the document box (==3)
+ // skip empty or status off categories
+ } else {
$content .= '<a class="' . $new_style . '" href="' . zen_href_link(FILENAME_DEFAULT, $box_categories_array[$i]['path']) . '">';
if ($box_categories_array[$i]['current']) {
| [RecordCount,Execute] | Zenue des categories This function returns a string with the content of the category box. | In this case, this skip could be replaced with `continue` |
@@ -1,14 +1,12 @@
import logging
-import os
import socket
-import subprocess
-import sys
-import tempfile
import time
+from django.conf import settings
+from django.core.cache import parse_backend_uri
from django.core.management.base import BaseCommand
-import redisutils
+# import redisutils
import redis as redislib
log = logging.getLogger('z.redis')
| [cleanup->[file_keys],vacuum->[keys->[keys],keys],Command->[handle->[vacuum,cleanup]]] | This function returns a generator that yields the list of keys in the system. Yields the buffer of the unique items in the system. | this looks like it can be deleted |
@@ -324,6 +324,7 @@ jhipster:
key: <%= rememberMeKey %>
<%_ } _%>
mail: # specific JHipster mail property, for standard properties see MailProperties
+ enabled: false
from: <%= baseName %>@localhost
base-url: http://my-server-url-to-change # Modify according to your server's URL
metrics: # DropWizard Metrics configuration, used by MetricsConfiguration
| [No CFG could be retrieved] | The configuration of the nagios object. This is a different keystore in production. | please add docs to the website as this is a change in behavior and people might be confused why emails are disabled |
@@ -66,11 +66,11 @@ Depending on the amount of segments that need compaction, it may take a while,
so consider using the ``--progress`` option.
A segment is compacted if the amount of saved space is above the percentage value
-given by the ``--threshold`` option. If ommitted, a threshold of 10% is used.
+given by the ``--threshold`` option. If omitted, a threshold of 10% is used.
When using ``--verbose``, borg will output an estimate of the freed space.
After upgrading borg (server) to 1.2+, you can use ``borg compact --cleanup-commits``
to clean up the numerous 17byte commit-only segments that borg 1.1 did not clean up
due to a bug. It is enough to do that once per repository.
-See :ref:`separate_compaction` in Additional Notes for more details.
\ No newline at end of file
+See :ref:`separate_compaction` in Additional Notes for more details.
| [No CFG could be retrieved] | This is a hack to make sure that the borg compacting is done at the same. | this is also auto-generated, see line 1 in this file. also check archiver.py. |
@@ -62,6 +62,7 @@ reintegrate_inflight_io(void *data)
NULL);
return 0;
}
+#endif
static void
reintegrate_with_inflight_io(test_arg_t *arg, daos_obj_id_t *oid,
| [No CFG could be retrieved] | function to create object of type n_r Create a new object with the specified rank and target. | (style) trailing whitespace |
@@ -580,6 +580,9 @@ import zeppelin from '../../../zeppelin';
$scope.exportToDSV = function(delimiter) {
var dsv = '';
+ var dateFinished = moment(paragraph.dateFinished).format('YYYY-MM-DD hh:mm:ss A');
+ var exportedFileName = paragraph.title ? paragraph.title + '_' + dateFinished : 'Exported data_' + dateFinished;
+
for (var titleIndex in tableData.columns) {
dsv += tableData.columns[titleIndex].name + delimiter;
}
| [No CFG could be retrieved] | copy common setting for vizId and copy values for vizId This function is used to resize the Viz instance. | nit: shell we make file name starts with lower case? `data_` instead of `Exported data_`, what do you think? |
@@ -317,7 +317,12 @@ export class AppStorage {
}
}
if (data !== null) {
- const realm = await this.getRealm();
+ let realm;
+ try {
+ realm = await this.getRealm();
+ } catch (error) {
+ alert(error.message);
+ }
data = JSON.parse(data);
if (!data.wallets) return false;
const wallets = data.wallets;
| [No CFG could be retrieved] | Load all wallets and a specific from storage. get the unserialized version of the wallet. | Why realm became optional? |
@@ -86,7 +86,9 @@ func NewMetricSet(base mb.BaseMetricSet) (*MetricSet, error) {
}
// Get IAM account name
- svcIam := iam.New(awsConfig)
+ awsConfig.Region = "us-east-1"
+ svcIam := iam.New(awscommon.EnrichAWSConfigWithEndpoint(
+ config.AWSConfig.Endpoint, "iam", "", awsConfig))
req := svcIam.ListAccountAliasesRequest(&iam.ListAccountAliasesInput{})
output, err := req.Send(context.TODO())
if err != nil {
| [AddModule,Warn,Logger,GetAWSCredentials,DescribeRegionsRequest,New,GetCallerIdentityRequest,Module,Send,Wrap,ListAccountAliasesRequest,TODO,UnpackConfig,Retrieve,Put] | NewModule creates a new module that can be used to create a new metricset. Construct MetricSet with a full regions list. | The hardcoded region here works well in global regions but not for China regions. |
@@ -265,6 +265,7 @@ class Openfoam(Package):
version('develop', branch='develop', submodules='True')
version('master', branch='master', submodules='True')
+ version('1912_200506', sha256='831a39ff56e268e88374d0a3922479fd80260683e141e51980242cc281484121')
version('1912_200403', sha256='1de8f4ddd39722b75f6b01ace9f1ba727b53dd999d1cd2b344a8c677ac2db4c0')
version('1912', sha256='437feadf075419290aa8bf461673b723a60dc39525b23322850fb58cb48548f2')
version('1906_200312', sha256='f75645151ed5d8c5da592d307480979fe580a25627cc0c9718ef370211577594')
| [Openfoam->[install_write_location->[rewrite_environ_files],patch->[add_extra_files,rewrite_environ_files],setup_dependent_run_environment->[setup_run_environment],install->[install,install_write_location],configure->[foam_add_lib,foam_add_path,write_environ,pkglib,rewrite_environ_files,mplib_content],setup_dependent_build_environment->[setup_run_environment],setup_run_environment->[setup_minimal_environment]],OpenfoamArch->[has_rule->[_rule_directory],create_rules->[_rule_directory,mplib_content]],write_environ->[_write_environ_file],_write_environ_file->[_write_environ_entries],mplib_content->[pkglib],_write_environ_entries->[_write_environ_entries]] | An example of how to use OpenFOAM s version of the sequence number. - - - - - - - - - - - - - - - - - -. | Apologies but our flake rules have changed and flake now thinks this is problematic. IMO we should figure out how to allow this since I think it looks better than what flake would be ok with (e.g. move this paren to the end of the previous line like `3e5d61')` but in the meantime if you are able to update this it will get it past our CI. |
@@ -42,7 +42,7 @@ def _process() -> bool:
Trade.session.add(trade)
state_changed = True
else:
- logging.info('Got no buy signal...')
+ logger.info('Got no buy signal...')
except ValueError:
logger.exception('Unable to create trade')
| [cleanup->[cleanup],init->[init],main->[init,_process],_process->[_process],handle_trade->[execute_sell,should_sell],create_trade->[get_target_bid],main] | Checks if a new trade has been created or closed. | I think now would be a good chance to change this to something more informative. Like, "Checked all whitelisted currencies. Found no suitable entry positions for buying. Will keep looking." |
@@ -52,8 +52,8 @@ def main():
else:
sys.exit("Error: did not expect more than one argument")
- paths = [binpath]
- sys.stderr.write("prepending %s to PATH\n" % binpath)
+ paths = binpath
+ sys.stderr.write("prepending %s to PATH\n" % ', '.join(binpath))
elif sys.argv[1] == '..deactivate':
if len(sys.argv) != 2:
| [binpath_from_arg->[prefix_from_arg],main->[binpath_from_arg,help],main] | Entry point for the command line interface. This function checks if the user provided a environment and if so checks if it has a con. | Maybe `' and '.join` or `os.pathsep.join`. |
@@ -545,7 +545,11 @@ def main():
def execbuffer(buf):
try:
- ret, ret_err = pyb.exec_raw(buf, timeout=None, data_consumer=stdout_write_bytes)
+ if args.no_follow:
+ pyb.exec_raw_no_follow(buf)
+ ret_err = None
+ else:
+ ret, ret_err = pyb.exec_raw(buf, timeout=None, data_consumer=stdout_write_bytes)
except PyboardError as er:
print(er)
pyb.close()
| [ProcessToSerial->[write->[write],read->[read]],execfile->[exit_raw_repl,enter_raw_repl,stdout_write_bytes,execfile,Pyboard,close],main->[execbuffer->[exec_raw,stdout_write_bytes,close,exit_raw_repl],exec_,execbuffer,exit_raw_repl,enter_raw_repl,stdout_write_bytes,follow,Pyboard,close,filesystem_command,read],Pyboard->[exec_->[exec_raw,PyboardError],fs_mkdir->[exec_],exit_raw_repl->[write],enter_raw_repl->[inWaiting,read,write,read_until,PyboardError],fs_get->[eval,exec_,write],fs_put->[exec_,read],fs_rm->[exec_],follow->[read_until,PyboardError],execfile->[exec_,read],__init__->[ProcessToSerial,write,TelnetToSerial,PyboardError,ProcessPtyToTerminal],exec_raw->[follow,exec_raw_no_follow],read_until->[inWaiting,read],fs_ls->[exec_],close->[close],fs_cat->[exec_],fs_rmdir->[exec_],exec_raw_no_follow->[read,write,read_until,PyboardError]],TelnetToSerial->[close->[close],write->[write],__init__->[PyboardError]],filesystem_command->[fname_cp_dest,close,fname_remote,exit_raw_repl],ProcessPtyToTerminal->[inWaiting->[inWaiting],write->[write],read->[read],__init__->[close]],main] | Main entry point for the command line interface. This function is called from the command line. It runs the command and if the command is. | Using `no_follow` here implements more of a "detach early" operation: in normal situations the script is piped to the board, run, and then pyboard.py waits for it to finish running. The `no_follow` option here is just forcing it to do the first part of that sequence, ie download script and run, then detach. So it might be better to call the option `--detach-early`.... but that's possibly confusing and in the end `--no-follow` is more logical because it corresponds to `exec_raw` vs `exec_raw_no_follow`. In summary I'm ok with this part of the PR. |
@@ -43,6 +43,11 @@ namespace System.Text.Json
JsonTokenType tokenType = reader.TokenType;
+ if (options.ReferenceHandling.ShouldReadPreservedReferences())
+ {
+ CheckValidTokenAfterMetadataValues(ref readStack, tokenType);
+ }
+
if (JsonHelpers.IsInRangeInclusive(tokenType, JsonTokenType.String, JsonTokenType.False))
{
Debug.Assert(tokenType == JsonTokenType.String || tokenType == JsonTokenType.Number || tokenType == JsonTokenType.True || tokenType == JsonTokenType.False);
| [JsonSerializer->[HandleObjectAsValue->[Read,TokenType,IsFinalBlock,TrySkip,HandleValue,AggressiveInlining,ReadAhead,BytesConsumed,Assert,Slice],GetUnescapedString->[Clear,Rent,StackallocThreshold,Return,ToArray,Unescape,Length],ReadCore->[StartObject,SkipProperty,HandleStartDictionary,ExceptionSourceValueToRethrowAsJsonException,False,ReadAhead,Source,HandleStartObject,Drain,IsInRangeInclusive,AddExceptionInformation,IsProcessingDictionary,BytesConsumed,Null,Assert,HandleEndArray,EndArray,StartArray,ReThrowWithPath,HandlePropertyName,TokenType,HandleValue,HandleStartArray,HandleEndDictionary,EndProperty,IsProcessingValue,CurrentState,HandleEndObject,PropertyName,Read,HandleNull,HandleObjectAsValue,String,Pop,EndObject,Push,True,Number]]] | Reads a core object. Checks if a node is found in the read stack and if so checks if it is a. | Not sure how yet... but we should consider how this check can be more targeted (in applicable `HandleX` methods rather than on the main loop). |
@@ -53,7 +53,8 @@ func newProcessorPipeline(
localProcessors = makeClientProcessors(config)
)
- needsCopy := global.alwaysCopy || localProcessors != nil || global.processors != nil
+ // needsCopy := global.alwaysCopy || localProcessors != nil || global.processors != nil
+ needsCopy := true
if !config.SkipNormalization {
// setup 1: generalize/normalize output (P)
| [Run->[fn,Run,Debug],String->[Sprintf,Join,String],Encode,IsDebug,Unlock,MergeFields,New,DeepUpdate,AddTags,ConvertToGenericEvent,All,Lock,Clone,add,Get,MapStrUnion,Err,Debug] | newProcessorPipeline prepares the processor pipeline for the given event object. Clone the fields and event metadata. | This looks weird. Is the intention to always copy? Or can the `needsCopy` value change to false in other parts of the code? |
@@ -116,6 +116,8 @@ def pandas_input_fn(x,
features = dict(zip(list(x.columns), features))
if y is not None:
target = features.pop(target_column)
+ # undo modification to dictionary
+ del x[target_column]
return features, target
return features
return input_fn
| [pandas_input_fn->[input_fn->[list,dict,dequeue_many,dequeue_up_to,pop,len,_enqueue_data,zip],TypeError,array_equal,max,len,copy,format,isinstance,ValueError]] | Returns input function that will feed pandas DataFrame into the model. Input function for picking a single missing node. | Line 79 of this file has x=x.copy(). So, same question: Do you really need this line? |
@@ -214,6 +214,10 @@ export class AccessService {
user.warn(TAG, 'Experiment "amp-access-server" is not enabled.');
type = AccessType.CLIENT;
}
+ if (type == AccessType.CLIENT && this.isServerEnabled_) {
+ user.info(TAG, 'Forcing access type: SERVER');
+ type = AccessType.SERVER;
+ }
return type;
}
| [No CFG could be retrieved] | Creates an adapter for the given access type. The login map is a map of URL - > URL pairs. | This kind of seems big for this PR. I'd not do it now, since it will start encrypting `type=client` reader IDs as well. |
@@ -0,0 +1,11 @@
+package games.strategy.engine.framework.map.download;
+
+interface DownloadListener {
+ void downloadStarted(DownloadFileDescription download);
+
+ void downloadUpdated(DownloadFileDescription download, long bytesReceived);
+
+ // TODO: possibly need to pass a second enum parameter describing the "reason" the download stopped
+ // (e.g. DONE, CANCELLED, etc.)
+ void downloadStopped(DownloadFileDescription download);
+}
\ No newline at end of file
| [No CFG could be retrieved] | No Summary Found. | Please add the EOF newline I forgot. |
@@ -318,6 +318,15 @@ public class WorkUnit extends State {
return value;
}
+ @Override
+ public String getProp(String key, String def) {
+ String value = super.getProp(key);
+ if (value == null) {
+ value = this.extract.getProp(key, def);
+ }
+ return value;
+ }
+
/**
* Set the low watermark of this {@link WorkUnit}.
*
| [WorkUnit->[readFields->[readFields],contains->[contains],copyOf->[WorkUnit],createEmpty->[WorkUnit],equals->[equals],write->[write],create->[WorkUnit],hashCode->[hashCode],getProp->[getProp],getLowWatermark->[getLowWatermark],getExpectedHighWatermark->[getExpectedHighWatermark]]] | Get the property. | getProp(String key) can call this one with null for default. |
@@ -30,11 +30,11 @@ class Sz(AutotoolsPackage):
"""Error-bounded Lossy Compressor for HPC Data."""
homepage = "https://collab.cels.anl.gov/display/ESR/SZ"
- url = "https://github.com/disheng222/SZ/archive/v1.4.9.2.tar.gz"
+ url = "https://github.com/disheng222/SZ/archive/v1.4.11.0.tar.gz"
version('develop', git='https://github.com/disheng222/SZ.git',
branch='master')
- version('1.4.9.2', '028ce90165b7a4c4051d4c0189f193c0')
+ version('1.4.11.0', '10dee28b3503821579ce35a50e352cc6')
variant('fortran', default=False,
description='Enable fortran compilation')
| [Sz->[variant,version]] | Configure the command line arguments for the . | Keep the old version! |
@@ -18,10 +18,16 @@ import com.amazonaws.services.s3.model.AmazonS3Exception;
import com.amazonaws.services.s3.model.CannedAccessControlList;
import com.amazonaws.services.s3.model.GetObjectMetadataRequest;
import com.amazonaws.services.s3.model.GetObjectRequest;
+import com.amazonaws.services.s3.model.ListObjectsRequest;
+import com.amazonaws.services.s3.model.ObjectListing;
import com.amazonaws.services.s3.model.ObjectMetadata;
import com.amazonaws.services.s3.model.PutObjectRequest;
import com.amazonaws.services.s3.model.PutObjectResult;
import com.amazonaws.services.s3.model.S3Object;
+import com.amazonaws.services.s3.model.S3ObjectSummary;
+import com.amazonaws.services.s3.model.StorageClass;
+
+import java.util.Date;
import static java.net.HttpURLConnection.HTTP_OK;
| [MockAmazonS3->[getObjectMetadata->[setStatusCode,AmazonS3Exception],getObject->[setStatusCode,AmazonS3Exception],putObject->[getCannedAcl,PutObjectResult]]] | Creates a mock object from an object in the system. This method is overridden to provide additional information for the object metadata. | Let's inline these since they are only used in one place (and the usage is self-descriptive) |
@@ -55,9 +55,12 @@ public final class WebApplicationServiceFactory extends AbstractServiceFactory<W
final String id = cleanupUrl(serviceToUse);
final String artifactId = request.getParameter(CasProtocolConstants.PARAMETER_TICKET);
- return new SimpleWebApplicationServiceImpl(id, serviceToUse,
- artifactId, "POST".equalsIgnoreCase(method) ? Response.ResponseType.POST
- : Response.ResponseType.REDIRECT);
+ final Response.ResponseType type = "POST".equalsIgnoreCase(method) ? Response.ResponseType.POST
+ : Response.ResponseType.REDIRECT;
+
+ final WebApplicationService webApplicationService = new SimpleWebApplicationServiceImpl(id, serviceToUse,
+ artifactId, new WebApplicationServiceResponseBuilder(type));
+ return webApplicationService;
}
@Override
| [WebApplicationServiceFactory->[createService->[getAttribute,hasText,equalsIgnoreCase,getParameter,cleanupUrl,SimpleWebApplicationServiceImpl]]] | Create a WebApplicationService based on the request parameters. | Do we need a new `ResponseBuilder` for each service? I tend to think the response builder is a singleton (one by type). |
@@ -7,8 +7,15 @@ d_echo('Ironware Mempool'."\n");
if (str_contains($device['sysDescr'], array('NetIron', 'MLX', 'CER')) === false) {
echo 'Ironware Dynamic: ';
$mempool['total'] = snmp_get($device, 'snAgGblDynMemTotal.0', '-OvQ', 'FOUNDRY-SN-AGENT-MIB');
+ if ($mempool['total']<0) {
+ $mempool['total'] =+ 4294967296 ; // signed vs unsigned snmp output
+ }
$mempool['free'] = snmp_get($device, 'snAgGblDynMemFree.0', '-OvQ', 'FOUNDRY-SN-AGENT-MIB');
+ if ($mempool['free']<0) {
+ $mempool['free'] =+ 4294967296 ; // signed vs unsigned snmp output
+ }
$mempool['used'] = ($mempool['total'] - $mempool['free']);
+ d_echo($mempool);
} //end_if
else {
| [No CFG could be retrieved] | Gets the memory usage of an OID. end region region region region region region region region region region region region region region region region region. | Can you pop spaces around the `<` in this + the free one below. |
@@ -820,6 +820,9 @@ class NestedArray(Array):
stride *= i
return tuple(reversed(strides))
+ @property
+ def key(self):
+ return self.dtype, self.shape
class UniTuple(IterableType):
| [WeakType->[__hash__->[__hash__]],Function->[extend->[extend]],Tuple->[coerce->[Tuple]],NoneType->[coerce->[Optional]],Array->[copy->[Array],post_init->[copy]],Optional->[coerce->[Optional]],UniTuple->[coerce->[UniTuple]],Buffer->[__init__->[ArrayIterator]],Dispatcher->[overloaded->[_get_object],__init__->[_store_object]],_TypeMetaclass->[__call__->[__call__,_autoincr]],NoneType,Boolean,Complex,RangeType,Dummy,Integer,RangeIteratorType,Opaque,Phantom,Slice3Type,Float] | Returns tuple of tuples of size and count where each tuple is a tuple of size and count. | Sorry for nitpicking, but this line is slightly misindented :-) Otherwise, patch looks good to me. |
@@ -49,7 +49,7 @@ class Photo {
public function __construct($data, $type=null) {
$this->imagick = class_exists('Imagick');
- $this->types = $this->supportedTypes();
+ $this->types = static::supportedTypes();
if (!array_key_exists($type, $this->types)){
$type='image/jpeg';
}
| [Photo->[orient->[is_imagick,rotate,flip,getType,is_valid],scaleImage->[scaleImage,is_imagick,getHeight,getWidth,is_valid],load_data->[get_FormatsMap,is_imagick],saveImage->[is_valid],__construct->[supportedTypes],rotate->[is_imagick,is_valid],store->[getWidth,getHeight,imageString,getType],getHeight->[is_imagick,is_valid],getExt->[is_valid,getType],flip->[is_imagick,is_valid],imageString->[is_imagick,is_valid,getType],getType->[is_valid],getImage->[is_imagick,is_valid],scaleImageSquare->[scaleImage,is_imagick,is_valid],cropImage->[scaleImage,is_imagick,is_valid,cropImage],scaleImageUp->[scaleImage,is_imagick,getHeight,getWidth,is_valid],getWidth->[is_imagick,is_valid],is_valid->[is_imagick]],import_profile_photo->[scaleImage,store,getExt,scaleImageSquare,is_valid],store_photo->[orient,scaleImage,store,getHeight,getExt,cropImage,getWidth,is_valid]] | Constructor for Image object. | I would say this you need to say `self::` instead of `static::` (`$this->types = self::supportedTypes();`) |
@@ -316,6 +316,10 @@ func (r *OCIRuntime) createOCIContainer(ctr *Container, cgroupParent string, res
cmd.Env = append(r.conmonEnv, fmt.Sprintf("_OCI_SYNCPIPE=%d", 3))
cmd.Env = append(cmd.Env, fmt.Sprintf("_OCI_STARTPIPE=%d", 4))
cmd.Env = append(cmd.Env, fmt.Sprintf("XDG_RUNTIME_DIR=%s", runtimeDir))
+ cmd.Env = append(cmd.Env, fmt.Sprintf("_LIBPOD_USERNS_CONFIGURED=%s", os.Getenv("_LIBPOD_USERNS_CONFIGURED")))
+ cmd.Env = append(cmd.Env, fmt.Sprintf("_LIBPOD_ROOTLESS_UID=%s", os.Getenv("_LIBPOD_ROOTLESS_UID")))
+ cmd.Env = append(cmd.Env, fmt.Sprintf("HOME=%s", os.Getenv("HOME")))
+ cmd.Env = append(cmd.Env, fmt.Sprintf("XDG_RUNTIME_DIR=%s", runtimeDir))
if r.reservePorts {
ports, err := bindPorts(ctr.config.PortMappings)
| [createOCIContainer->[IsRootless,CurrentLabel,Close,SetProcessLabel,LookupEnv,deleteContainer,LogPath,LockOSThread,GetRootlessRuntimeDir,Start,Errorf,After,Wait,Debugf,ReadBytes,ID,Wrapf,Join,moveConmonToCgroup,UnlockOSThread,Get,Files,CheckpointPath,Command,Write,NewReader,Sprintf,Unmarshal,NewContext,GetEnabled,bundlePath,GetLevel,String,Pipe,WithFields],killContainer->[ID,Wrapf,Sprintf,GetRootlessRuntimeDir,ExecCmdWithStdStreams,Debugf],stopContainer->[ID,Wrapf,IsRootless,killContainer,Sprintf,Kill,Warnf,Duration,GetRootlessRuntimeDir,ExecCmdWithStdStreams,Debugf],pauseContainer->[ExecCmdWithStdStreams,Sprintf,ID,GetRootlessRuntimeDir],unpauseContainer->[ExecCmdWithStdStreams,Sprintf,ID,GetRootlessRuntimeDir],updateContainerStatus->[Now,Close,StderrPipe,removeConmonFiles,IsNotExist,ExponentialBackoff,Stat,exitFilePath,GetRootlessRuntimeDir,Start,Errorf,Wait,handleExitFile,ID,NewBuffer,Wrapf,Contains,StdoutPipe,NewDecoder,Command,Sprintf,Decode,ReadAll],startContainer->[ID,Sprintf,Now,GetRootlessRuntimeDir,ExecCmdWithStdStreams],execStopContainer->[ID,Wrapf,Sprintf,Kill,Warnf,GetRootlessRuntimeDir,Duration,ExecCmdWithStdStreams,Debugf],checkpointContainer->[ID,bundlePath,ExecCmdWithStdStreams,CheckpointPath,Debugf],execContainer->[LogPath,ID,Wrapf,Sprintf,GetRootlessRuntimeDir,Start,execPidPath,Command,Debugf],deleteContainer->[ID,ExecCmd],ID,Wrapf,Join,ListenTCP,Sprintf,ListenUDP,Kill,File,ResolveTCPAddr,ResolveUDPAddr,Errorf,After,Sleep,MkdirAll,IsExist] | createOCIContainer creates an OCI container start conmon if necessary BindPorts binds the ports to the container and then starts the SELinux context This function is called by the init process when a container is created This function is used to read the container state from the channel and create it if necessary. | Should we check if they are set within the environment before setting them for conmon? Is there any difference in the code between +LIBPOD+USERNS_CONFIGURED="" versus not being set at all? Or for any of the others? |
@@ -78,7 +78,10 @@ func (a *apiServer) Extract(request *admin.ExtractRequest, extractServer admin.A
defer func(start time.Time) { a.Log(request, nil, retErr, time.Since(start)) }(time.Now())
ctx := extractServer.Context()
pachClient := a.getPachClient().WithCtx(ctx)
- writeOp := extractServer.Send
+ writeOp := func(op *admin.Op) error {
+ fmt.Println(op)
+ return extractServer.Send(op)
+ }
if request.URL != "" {
url, err := obj.ParseURL(request.URL)
if err != nil {
| [Restore->[WithCtx,Context,SendAndClose,Read,getPachClient,Write,applyOp,NewReader,Now,ParseURL,NewClientFromURLAndSecret,Reader,Log,Recv,Errorf,Since,PutObject],getPachClient->[NewFromAddress,Do,Sprintf],applyOp->[NewBranch,SanitizeName,IsAlreadyExistError,CreatePipeline,BuildCommit,TagObject,Ctx,Errorf,ScrubGRPC,CreateRepo,CreateBranch],Extract->[Context,GetObject,NewBufferedWriter,ListObject,Now,ParseURL,Close,NewClientFromURLAndSecret,Writer,ListRepo,ListPipeline,Errorf,Since,NewWriter,ListCommit,Log,ListBranch,WithCtx,Write,getPachClient,ListTag],ExtractPipeline->[WithCtx,getPachClient,Now,Log,InspectPipeline,Since],Read->[Write,Read,Reset,Errorf,Len,ScrubGRPC,Recv],MustCompile,SanitizeName,VisitInput,NewCommit] | Extract - extracts objects from a remote object This function is used to read all objects in the cluster. write all the operations that need to be sent to the client. | is this debugging code or is this intentionally left in? |
@@ -132,10 +132,13 @@ public class OMKeyCommitRequest extends OMKeyRequest {
keyName, IAccessAuthorizer.ACLType.WRITE,
commitKeyRequest.getClientID());
- List<OmKeyLocationInfo> locationInfoList = commitKeyArgs
- .getKeyLocationsList().stream()
- .map(OmKeyLocationInfo::getFromProtobuf)
- .collect(Collectors.toList());
+ List<OmKeyLocationInfo> locationInfoList = new ArrayList<>();
+ List<KeyLocation> keyLocations = commitKeyArgs
+ .getKeyLocationsList();
+
+ for (KeyLocation keyLocation : keyLocations) {
+ locationInfoList.add(OmKeyLocationInfo.getFromProtobuf(keyLocation));
+ }
bucketLockAcquired = omMetadataManager.getLock().acquireLock(BUCKET_LOCK,
volumeName, bucketName);
| [OMKeyCommitRequest->[preExecute->[build,getKeyArgs,checkNotNull,getCommitKeyRequest,now,setModificationTime],validateAndUpdateCache->[getUserInfo,size,incNumKeyCommits,getClientID,incNumKeyCommitFails,OMReplayException,getOmRequest,getVolumeName,getMetrics,OMKeyCommitResponse,of,isDBOperationNeeded,getOpenKey,createErrorOMResponse,getBucketName,auditLog,checkKeyAclsInOpenKeyTable,getOzoneKey,setDataSize,setModificationTime,releaseLock,getKeyName,setUpdateID,debug,isRatisEnabled,error,getIfExist,getCommitKeyRequest,acquireLock,absent,getDataSize,incNumKeys,buildKeyArgsAuditMap,isReplay,OMException,getAuditLogger,buildAuditMessage,getModificationTime,toList,getKeyArgs,validateBucketAndVolume,addCacheEntry,updateLocationInfoList,get,build,collect,createReplayOMResponse,getMetadataManager,addResponseToDoubleBuffer,getOMResponseBuilder],getLogger]] | This method checks if the given OzoneKey already exists in the database and updates the cache This method checks if the key is in the open key table and if it is in the This method is called when a response from the OOM layer is received. | The temporary variable `keyLocations ` can be removed. |
@@ -2187,14 +2187,8 @@ public class NetworkModelImpl extends ManagerBase implements NetworkModel, Confi
@Override
public void checkIp6Parameters(String startIPv6, String endIPv6, String ip6Gateway, String ip6Cidr) throws InvalidParameterValueException {
- if (!NetUtils.isValidIp6(startIPv6)) {
- throw new InvalidParameterValueException("Invalid format for the startIPv6 parameter");
- }
- if (!NetUtils.isValidIp6(endIPv6)) {
- throw new InvalidParameterValueException("Invalid format for the endIPv6 parameter");
- }
- if (!(ip6Gateway != null && ip6Cidr != null)) {
+ if (org.apache.commons.lang3.StringUtils.isBlank(ip6Gateway) || org.apache.commons.lang3.StringUtils.isBlank(ip6Cidr)) {
throw new InvalidParameterValueException("ip6Gateway and ip6Cidr should be defined when startIPv6/endIPv6 are passed in");
}
| [NetworkModelImpl->[getUserDataUpdateProvider->[getElementImplementingProvider],areServicesSupportedInNetwork->[areServicesSupportedInNetwork],getNetworkRate->[getNetwork],getNicProfile->[getNetworkTag,getNetworkRate,isSecurityGroupSupportedInNetwork],isSecurityGroupSupportedInNetwork->[findPhysicalNetworkId],isServiceEnabledInNetwork->[isProviderEnabledInPhysicalNetwork,areServicesSupportedInNetwork],getDefaultUniqueProviderForService->[listSupportedNetworkServiceProviders],areServicesEnabledInZone->[isProviderEnabledInPhysicalNetwork,findPhysicalNetworkId],getProviderToIpList->[getElementImplementingProvider,getProviderServicesMap],canElementEnableIndividualServices->[getElementImplementingProvider],checkCapabilityForProvider->[getElementImplementingProvider],canProviderSupportServices->[getElementImplementingProvider],listNetworkOfferingsForUpgrade->[areServicesSupportedByNetworkOffering],getDefaultStorageTrafficLabel->[getDefaultPhysicalNetworkByZoneAndTrafficType],getDefaultManagementTrafficLabel->[getDefaultPhysicalNetworkByZoneAndTrafficType],getSourceNatIpAddressForGuestNetwork->[listPublicIpsAssignedToGuestNtwk],getNetworkOfferingServiceCapabilities->[getElementImplementingProvider],getIpInNetwork->[getNicInNetwork],getNtwkOffDetails->[getNtwkOffDetails],isProviderEnabledInZone->[isProviderEnabledInPhysicalNetwork],getNonGuestNetworkPhysicalNetworkId->[getPhysicalNetworkId],getElementServices->[getElementImplementingProvider],canIpUsedForNonConserveService->[getIpToServices],checkIpForService->[canIpUsedForNonConserveService,canIpUsedForService],areServicesSupportedByNetworkOffering->[areServicesSupportedByNetworkOffering],getIpToServices->[getPublicIpPurposeInRules],getNetworkServiceCapabilities->[getElementImplementingProvider],isProviderForNetwork->[isProviderForNetwork],verifyDisabledConfigDriveEntriesOnEnabledZones->[addDisabledConfigDriveEntriesOnZone],isNetworkReadyForGc->[networkIsConfiguredForExternalNetworking,getNetwork],getPhysicalNetworkId->[findPhysicalNetworkId,getPhysicalNetworkId],canIpUsedForService->[getIpToServices,getServiceProvidersMap,getElementImplementingProvider],getNetworkCapabilities->[getElementImplementingProvider],providerSupportsCapability->[getElementImplementingProvider],getNetworkTag->[getNetworkTag,findPhysicalNetworkId],isNetworkAvailableInDomain->[getNetwork],isProviderForNetworkOffering->[isProviderForNetworkOffering]]] | check if the given ip6 parameters are valid and if not throw an InvalidParameterValueException. | can you encapsulate this in our local proxy to the StringUtils, so migration efforts will be located in a single file? |
@@ -750,8 +750,10 @@ void tool_change(const uint8_t tmp_extruder, const float fr_mm_s/*=0.0*/, bool n
const float old_feedrate_mm_s = fr_mm_s > 0.0 ? fr_mm_s : feedrate_mm_s;
feedrate_mm_s = fr_mm_s > 0.0 ? fr_mm_s : XY_PROBE_FEEDRATE_MM_S;
- #if HAS_SOFTWARE_ENDSTOPS && ENABLED(DUAL_X_CARRIAGE)
+ #if HAS_SOFTWARE_ENDSTOPS
update_software_endstops(X_AXIS, active_extruder, tmp_extruder);
+ update_software_endstops(Y_AXIS, active_extruder, tmp_extruder);
+ update_software_endstops(Z_AXIS, active_extruder, tmp_extruder);
#endif
set_destination_from_current();
| [No CFG could be retrieved] | - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -. | Do you suppose this is okay in all cases? |
@@ -374,13 +374,15 @@ func (o *ImportImageOptions) createImageImport() (*imageapi.ImageStream, *imagea
},
Spec: imageapi.ImageStreamImportSpec{Import: true},
}
+ insecureAnnotation := stream.Annotations[imageapi.InsecureRepositoryAnnotation]
+ insecure := o.Insecure || insecureAnnotation == "true"
if o.All {
isi.Spec.Repository = &imageapi.RepositoryImportSpec{
From: kapi.ObjectReference{
Kind: "DockerImage",
Name: from,
},
- ImportPolicy: imageapi.TagImportPolicy{Insecure: o.Insecure},
+ ImportPolicy: imageapi.TagImportPolicy{Insecure: insecure},
}
} else {
isi.Spec.Images = append(isi.Spec.Images, imageapi.ImageImportSpec{
| [createImageImport->[Get,FollowTagReference,Errorf,IsNotFound],Error->[Sprintf],waitForImport->[Stop,ResultChan,Watch,Errorf,Parse,OneTermEqualSelector],Validate->[UsageError,ParseDockerImageReference,Errorf],Run->[Describe,createImageImport,Fprintf,Fprintln,Join,Create,waitForImport,Fprint,Import,Errorf,IsZero,Update],Complete->[DefaultNamespace,ImageStreams,Clients],StringVar,Flags,Sprintf,Validate,CheckErr,Run,Complete,BoolVar] | createImageImport creates an image stream and an image stream import from the specified image stream. import all images in the image stream tag - tag - tag - tag - reference - from - image stream ImageImportSpec - ImageImportSpec for Docker images. | If a user passes `oc import-image --insecure=false`, should that take priority over the image stream? Why or why not? |
@@ -420,3 +420,7 @@ class Configuration(object):
file_parser_tuple = (fname, parser)
if file_parser_tuple not in self._modified_parsers:
self._modified_parsers.append(file_parser_tuple)
+
+ def __repr__(self):
+ # type: () -> str
+ return "{}({!r})".format(self.__class__.__name__, self._dictionary)
| [Configuration->[_load_file->[items],_get_environ_vars->[items],_iter_config_files->[get_configuration_files],_normalized_keys->[_normalize_name],_load_config_files->[items],items->[items],unset_value->[_disassemble_key,items],set_value->[_disassemble_key]]] | Mark a parser as modified. | There's more things that represent a Configuration object (eg. parsers, load_only etc) which we should probably include in the `repr`. |
@@ -31,7 +31,7 @@ func mkChunk(t require.TestingT, from model.Time, points int) chunk.Chunk {
metric := model.Metric{
model.MetricNameLabel: "foo",
}
- pc, err := promchunk.NewForEncoding(promchunk.DoubleDelta)
+ pc, err := promchunk.NewForEncoding(promchunk.Bigchunk)
require.NoError(t, err)
ts := from
for i := 0; i < points; i++ {
| [Time,Seek,reset,Itoa,EqualValues,NewForEncoding,NewChunk,At,Next,TimeFromUnix,False,NoError,Len,SampleValue,Err,Add,True] | TestChunkIter imports a chunk from a given time. Time and points. testSeek tests that t has the same values as ts and that t has the same value. | Shouldn't we test for each chunk encoding? |
@@ -26,6 +26,15 @@ export default function Sidebar( { settings, store, theme } ) {
return (
<Fragment>
<Fill name="YoastSidebar">
+ { <SidebarItem renderPriority={ 5 }>
+ <ThemeProvider theme={ theme }>
+ <StoreProvider store={ store }>
+ <div>
+ <SidebarSnippetPreviewButton />
+ </div>
+ </StoreProvider>
+ </ThemeProvider>
+ </SidebarItem> }
{ settings.isContentAnalysisActive && <SidebarItem renderPriority={ 10 }>
<ThemeProvider theme={ theme }>
<StoreProvider store={ store }>
| [No CFG could be retrieved] | The Sidebar component. The Sidebar class shows the show of the missing object. | Why this div? |
@@ -763,7 +763,7 @@ func TestCorruptEncryption(t *testing.T) {
t.Fatal(err)
}
_, _, err = Open(ciphertext, kr)
- if mm, ok := err.(ErrBadCiphertext); !ok {
+ if mm, ok := err.(ErrBadTag); !ok {
t.Fatalf("Got wrong error; wanted 'Bad Ciphertext' but got %v", err)
} else if int(mm) != 3 {
t.Fatalf("Wanted a failure in packet %d but got %d", 3, mm)
| [Box->[ToRawBoxKeyPointer],Unbox->[ToRawBoxKeyPointer],Precompute->[ToRawBoxKeyPointer,Precompute],GetPublicKey,makeIterable,insert,CreateEphemeralKey,ToKID] | TestCorruptEncryption tests that a message is encrypted with a Poly1305 tag. nanoname is the same as above but it s not the same as the original. | Note that in this approach, it's impossible to produce `ErrBadCiphertext` by corrupting the ciphertext after encryption, because the per-recipient authenticator now covers the entire ciphertext and always leads to an `ErrBadTag`. To get the `ErrBadCiphertext`, we have to corrupt the ciphertext before the authenticators are computed by the sender. Line 707 tests this by messing with the session key. |
@@ -129,6 +129,10 @@ func (o *Operation) ID() string {
return o.id
}
+func (o *Operation) Auditf(format string, args ...interface{}) {
+ o.Infof(format, args...)
+}
+
func (o *Operation) Infof(format string, args ...interface{}) {
o.Info(fmt.Sprintf(format, args...))
}
| [Fatal->[Fatalf,Fatal,header],Panic->[Panicf,header,Panic],Err->[Err],Info->[Info,header,Infof],Error->[Error,header,Errorf],Warn->[Warn,Warnf,header],String->[header],Debug->[Debugf,header,Debug],Debugf,header,newChild] | ID returns the ID of the operation. | Is there a design available for this work? I assume there's a reason to introduce `Auditf` as a distinct method (instead of just using `Infof`), but it's not clear from this PR. Understanding the long-term plan will help with review of this PR. Some initial thoughts: * Maybe this method should call `Infof` to avoid duplicate code (or maybe that's a silly suggestion given the long-term plan) * Maybe a `trace.BeginAndAudit` method (or just `trace.Audit`?) could be introduced that calls `trace.Begin` and `op.Auditf` to avoid the need to for two lines duplicating the operation name at each of these call sites |
@@ -18,6 +18,11 @@
-%>
const webpackConfig = require('../../../webpack/webpack.test.js');
+var ChromiumRevision = require('puppeteer/package.json').puppeteer.chromium_revision;
+var Downloader = require('puppeteer/utils/ChromiumDownloader');
+var revisionInfo = Downloader.revisionInfo(Downloader.currentPlatform(), ChromiumRevision);
+process.env.CHROMIUM_BIN = revisionInfo.executablePath;
+
const WATCH = process.argv.indexOf('--watch') > -1;
module.exports = (config) => {
| [No CFG could be retrieved] | A module that exports a single that can be found in the browser. Options for the karma reporters. | use `const` please |
@@ -16,8 +16,8 @@ return [
'help' => 'Amount of poller workers to spawn on this node.',
],
'poller_frequency' => [
- 'description' => 'Poller Frequency',
- 'help' => 'How often to to poll devices on this node. Warning! Changing this is not recommended. See docs for more info.',
+ 'description' => 'Poller Frequency (Warning!)',
+ 'help' => 'How often to to poll devices on this node. Warning! Changing this without fixing rrd files will break graphs. See docs for more info.',
],
'poller_down_retry' => [
'description' => 'Device Down Retry',
| [No CFG could be retrieved] | This function returns a list of all the properties of a given node. This function returns a list of services to run on this node. | "... to to ..." ? :D |
@@ -498,8 +498,10 @@ class ContainerProxy(object):
if indexing_directive is not None:
request_options["indexingDirective"] = indexing_directive
+ passthrough_args = {k: v for k, v in kwargs.items() if k != "enable_automatic_id_generation"}
+
result = self.client_connection.CreateItem(
- database_or_container_link=self.container_link, document=body, options=request_options, **kwargs
+ database_or_container_link=self.container_link, document=body, options=request_options, **passthrough_args
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
| [ContainerProxy->[read_item->[_set_partition_key,_get_document_link],is_system_key->[_get_properties],get_conflict->[_get_conflict_link,_set_partition_key],delete_item->[_set_partition_key,_get_document_link],replace_item->[_get_document_link],query_conflicts->[_set_partition_key],delete_conflict->[_get_conflict_link,_set_partition_key],query_items->[_set_partition_key],replace_throughput->[_get_properties],read_offer->[_get_properties]]] | Create a new item in the container. Create a new node in the database. | This line shouldn't be necessary as `enable_automatic_id_generation` has already been `popped` from the `kwargs` above. |
@@ -588,6 +588,13 @@
SERIAL_ECHOPGM("/256");
}
break;
+ case TMC_INTERPOLATE:
+ {
+ CHOPCONF_t r{0};
+ r.sr = st.CHOPCONF();
+ serialprint_truefalse(r.intpol);
+ }
+ break;
default: break;
}
}
| [No CFG could be retrieved] | Method to print out the status of the stepper. Private method for serial printing of the status of the TMC2208Stepper. | Why not just call `st.intpol()`? |
@@ -24,6 +24,10 @@ $graph_conf = array(
'svg' => 'svg',
),
),
+ array('name' => 'webui.graph_stacked',
+ 'descr' => 'Set display of stacked graphs',
+ 'type' => 'checkbox',
+ ),
);
$availability_map_conf = array(
| [No CFG could be retrieved] | This function refreshes the configuration of the web UI. Displays a panel showing the user s network network network network network network network network network network network. | You've used two variables now for this. I'd actually say `graphs.stacked` is better but I realise I used `webui.graph_type` for svg so I'd be fine with keeping this as well. |
@@ -38,6 +38,8 @@ describe('amp-inabox', () => {
installAmpdocServices(ampdoc);
const installedServicesByRegularAmp = installedServices.slice(0);
+ // The inabox mode does not need the loading indicator.
+ removeItem(installedServicesByRegularAmp, 'loadingIndicator');
installedServices = [];
installAmpdocServicesForInabox(ampdoc);
expect(installedServices).to.deep.equal(installedServicesByRegularAmp);
| [No CFG could be retrieved] | This method installs all services for the given AMP doc. | This seems fishy to me. I don't have the background here, but do you know why we are adding all the services by regular amp and then removing vs. only installing the correct ones to begin with? |
@@ -82,6 +82,9 @@ namespace System.Globalization
fixed (char* pSource = &MemoryMarshal.GetReference(source))
fixed (char* pValue = &MemoryMarshal.GetReference(value))
{
+ Debug.Assert(pSource != null);
+ Debug.Assert(pValue != null);
+
int ret = Interop.Kernel32.FindStringOrdinal(
dwFindStringOrdinalFlags,
pSource,
| [CompareInfo->[StartsWith->[FindString],LastIndexOfOrdinalCore->[FindStringOrdinal],IndexOfCore->[FindString],IndexOfOrdinalCore->[FindStringOrdinal],FindStringOrdinal->[FindStringOrdinal],LastIndexOfCore->[FindString],EndsWith->[FindString]]] | FindStringOrdinal - Find string ordinal in the given string. | >GetLastWin32Error [](start = 39, length = 17) I don't like checking the last error. Windows can change this and break us. I would like to depend on what the public doc says and for special cases we can handle it manually. |
@@ -2,6 +2,7 @@ import graphene
import graphene_django_optimizer as gql_optimizer
from django.core.exceptions import ValidationError
from graphene import relay
+from graphql_jwt.decorators import permission_required
from ...order import models
from ...order.models import FulfillmentStatus
| [OrderEvent->[resolve_lines->[OrderEventOrderLineObject]]] | A base class for all the fields that are needed to be sent to the order. Required fields uration and composed_id are required for the Fulfillment. | `from ..decorators import permission_required` |
@@ -149,7 +149,7 @@ class Command(BaseCommand):
send_mail(subject,
content,
- from_email='nobody@mzilla.org',
+ from_email='nobody@mozilla.org',
recipient_list=[recipient],
html_message=content,
reply_to=[recipient])
| [Command->[generate_report_html->[transform,join],mail_report->[info,send_mail],fetch_report_data->[append,get,read,fetchone,timedelta,cursor,open,fetchall,today,weekday,date,execute],handle->[info,mail_report,generate_report_html,fetch_report_data]],getLogger,join] | Sends a report to a recipient. | Typo drive-by fix. |
@@ -605,9 +605,6 @@ static int register_shortcut_event(lua_State *L)
// set up the accelerator path
dt_accel_connect_lua(tmp, g_cclosure_new(G_CALLBACK(shortcut_callback), tmp, closure_destroy));
- // free temporary buffer
- free(tmp);
-
return result;
}
| [dt_lua_event_trigger_wrapper->[dt_lua_event_trigger],int->[dt_lua_event_keyed_destroy,dt_lua_event_keyed_register],dt_lua_init_events->[dt_lua_event_add]] | register_shortcut_event registers the shortcut event in the lua stack and destroys the shortcut. | I see, `closure_destroy` will free `tmp`. |
@@ -609,7 +609,7 @@ class HelpfulArgumentParser(object):
"""
# Maps verbs/subcommands to the functions that implement them
- VERBS = {"auth": auth, "config_changes": config_changes,
+ VERBS = {"auth": obtaincert, "certonly": obtaincert, "config_changes": config_changes,
"install": install, "plugins": plugins_cmd,
"revoke": revoke, "rollback": rollback, "run": run}
| [_plugins_parsing->[add,add_group,add_plugin_args],_auth_from_domains->[_report_new_cert,_treat_as_renewal],revoke->[revoke,_determine_account],auth->[_auth_from_domains,_find_domains,_report_new_cert,choose_configurator_plugins,_init_le_client],_create_subparsers->[add,add_group,flag_default],setup_logging->[setup_log_file_handler],_treat_as_renewal->[_find_duplicative_certs],SilentParser->[add_argument->[add_argument]],HelpfulArgumentParser->[add_plugin_args->[add_group],parse_args->[parse_args],add->[add_argument],__init__->[SilentParser,flag_default]],install->[_init_le_client,_find_domains,choose_configurator_plugins],run->[_init_le_client,_auth_from_domains,_find_domains,choose_configurator_plugins],main->[setup_logging,prepare_and_parse_args],choose_configurator_plugins->[set_configurator,diagnose_configurator_problem],_paths_parser->[config_help,add,add_group,flag_default],_init_le_client->[_determine_account],prepare_and_parse_args->[HelpfulArgumentParser,flag_default,add,parse_args,config_help,add_group],rollback->[rollback],main] | Creates a class that wraps argparse. ArgumentParser and wraps it with a SilentArgumentParser. This function is called by the command line parser when the user has specified a command line argument. | nit: The alias `auth` is included here but not `everything`. I think we should have the same behavior for both aliases. The only behavior that I'm aware of that is affected here is whether you can get help using the alias (e.g. `letsencrypt --help auth`). I don't have a strong opinion, but I think aliases should not be included. The aliases are not shown in the help output and the distinction between `letsencrypt --help all` and `letsencrypt --help everything` is somewhat confusing. |
@@ -27,6 +27,7 @@ __all__ = [
'CenterNet',
'DetectionWithLandmarks',
'FaceBoxes',
+ 'Model',
'RetinaFace',
'SegmentationModel',
'SSD',
| [No CFG could be retrieved] | All of the classes that implement the ethernet interface. | Put `DeblurringModel` into common folder instead. |
@@ -47,5 +47,7 @@ define([
oneTimeWarning.geometryOutlines = 'Entity geometry outlines are unsupported on terrain. Outlines will be disabled. To enable outlines, disable geometry terrain clamping by explicitly setting height to 0.';
+ oneTimeWarning.geometryZIndex = 'Entity geometry with zIndex are unsupported when the entity is dynamic, or height or extrudedHeight are defined. zIndex will be ignored';
+
return oneTimeWarning;
});
| [No CFG could be retrieved] | The one - time warning is not supported on terrain. | Are dynamic entities still not supported? This message appears to only be used in one place and doesn't check anything about dynamic. |
@@ -71,7 +71,9 @@ export class VisibilityModel {
});
this.eventPromise_.then(() => {
- this.onTriggerObservable_.fire();
+ if (this.onTriggerObservable_) {
+ this.onTriggerObservable_.fire();
+ }
});
/** @private {!Array<!UnlistenDef>} */
| [No CFG could be retrieved] | Replies the properties of a single network network condition. The type of information is passed to the constructor. | How about we dev().warn in an else preferably with info that may help debug the problem (e.g. page url assuming it isn't already present) |
@@ -75,8 +75,16 @@ public final class MongoClientInstrumentation extends Instrumenter.Default {
public static class MongoClientAdvice {
@Advice.OnMethodEnter(suppress = Throwable.class)
- public static void injectTraceListener(@Advice.This final MongoClientOptions.Builder builder) {
- builder.addCommandListener(new TracingCommandListener());
+ public static void injectTraceListener(
+ @Advice.This final MongoClientOptions.Builder builder,
+ @Advice.FieldValue("commandListeners") final List<CommandListener> commandListeners) {
+ for (final CommandListener commandListener : commandListeners) {
+ if (commandListener instanceof TracingCommandListener) {
+ return;
+ }
+ }
+ final TracingCommandListener listener = new TracingCommandListener();
+ builder.addCommandListener(listener);
}
}
}
| [MongoClientInstrumentation->[typeMatcher->[declaresMethod,and,isPublic],MongoClientAdvice->[injectTraceListener->[addCommandListener,TracingCommandListener]],transformers->[singletonMap,and,takesArguments,getName]]] | Injects a trace listener into the MongoClientOptions. | Inline this variable |
@@ -35,6 +35,10 @@ h4 {
font-weight: 300;
}
+.container-fluid {
+ padding-top: 45px;
+}
+
<%_ if (clientTheme === 'none') { _%>
/* Increase contrast of links to get 100% on Lighthouse Accessability Audit. Override this color if you want to change the link color, or use a Bootswatch theme */
a {
| [No CFG could be retrieved] | Creates a new object. DETAILS - . | I'm not a CSS specialist, but is not a good idea to use px to set a padding value? |
@@ -97,8 +97,8 @@
].select(&:present?),
) %>
- <h2><%= t('doc_auth.instructions.privacy') %></h2>
- <p>
+ <h3><%= t('doc_auth.instructions.privacy') %></h3>
+ <p class="padding-bottom-2">
<%= t(
'doc_auth.info.privacy_html',
app_name: APP_NAME,
| [No CFG could be retrieved] | Renders the list of missing items. Displays a link to the user s IDV page that can be used to continue the authentication. | Poking around some of the other flow designs, I wonder if we could update the "Cancel" partial to have `margin-top-4` to achieve this same effect? cc @anniehirshman-gsa do you think this makes sense? In other words, using 2rem top margin wherever we have "Cancel" link at the bottom of a page, where currently it is 1rem by default. We already do this with the "Start Over or Cancel" variation. |
@@ -281,7 +281,7 @@ def cwt_morlet(X, sfreq, freqs, use_fft=True, n_cycles=7.0, zero_mean=False,
"""
mode = 'same'
# mode = "valid"
- n_signals, n_times = X.shape
+ n_signals, n_times = X[:, ::decim].shape
decim = int(decim)
# Precompute wavelets for given frequency range to save time
| [_cwt->[_centered],AverageTFR->[plot->[_preproc_tfr],__iadd__->[_check_compat],__isub__->[_check_compat],__add__->[_check_compat],__sub__->[_check_compat],plot_topo->[_preproc_tfr]],_induced_power_cwt->[morlet],_time_frequency->[_cwt],tfr_morlet->[AverageTFR,_induced_power_cwt,_get_data],tfr_multitaper->[_prepare_picks,AverageTFR,_induced_power_mtm,_get_data],_induced_power_mtm->[_dpss_wavelet],cwt->[_cwt],cwt_morlet->[morlet],read_tfrs->[AverageTFR],single_trial_power->[morlet,cwt],write_tfrs->[_prepare_write_tfr]] | Compute time - frequency decomposition with Morlet wavelets. | decim = int(decim) should be written before calling X[:, ::decim] |
@@ -134,7 +134,7 @@ public class TcpNioConnection extends TcpConnectionSupport {
@SuppressWarnings("unchecked")
public void send(Message<?> message) throws Exception {
- synchronized(this.getMapper()) {
+ synchronized(this.socketChannel) {
Object object = this.getMapper().fromMessage(message);
this.lastSend = System.currentTimeMillis();
try {
| [TcpNioConnection->[doClose->[close],doRead->[close,allocate],run->[isOpen],getPort->[getPort],isOpen->[isOpen],ChannelInputStream->[close->[close],read->[read]],readPacket->[doRead],allocate->[allocate],ChannelOutputStream->[close->[doClose],doWrite->[write]],convert->[dataAvailable]]] | This method sends a message to the channel. | Hi, Gary! Does it make sense to synchronize here at all? Thanks |
@@ -698,7 +698,7 @@ Spdp::handle_participant_data(DCPS::MessageId id,
if (from_relay) {
tport_->write_i(guid, iter->second.local_address_, SpdpTransport::SEND_RELAY);
} else {
- tport_->shorten_local_sender_delay_i();
+ tport_->zero_local_sender_delay_i ();
}
}
| [No CFG could be retrieved] | The first parties of a participant discovery are the participants that have been discovered. This method is called to populate the second part of the authentication process. | This could simply be `tport_->write_i(guid, iter->second.local_address_, SpdpTransport::SEND_LOCAL);` |
@@ -14,13 +14,16 @@
package api
import (
+ "fmt"
"io/ioutil"
"net/http"
"net/url"
"strings"
+ "go.uber.org/zap"
+
+ log "github.com/pingcap/log"
"github.com/pingcap/pd/server"
- log "github.com/sirupsen/logrus"
)
const (
| [ServeHTTP->[Header,Write,Join,Error,Do,ServeHTTP,IsLeader,Name,Close,Errorf,GetClientUrls,ParseUrls,Get,ReadAll,Set,WriteHeader,GetLeader],Add] | ServeHTTP handles a single HTTP request. ServeHTTP serves a custom reverse proxy for the given urls. | Remove empty line? |
@@ -396,10 +396,8 @@ class TestFloor(OpTest):
def test_check_output(self):
self.check_output()
- def test_check_grad(self):
- if self.dtype == np.float16:
- return
- self.check_grad(['X'], 'Out', max_relative_error=0.007)
+ # the gradient on floor, ceil, round is undefined.
+ # we return zero as gradient, but the numpy return nan
def init_dtype(self):
pass
| [TestLogSigmoid->[setUp->[init_dtype]],TestSin->[setUp->[init_dtype]],TestHardSigmoid->[setUp->[init_dtype]],TestHardShrink->[setUp->[init_dtype]],TestThresholdedRelu->[setUp->[init_dtype]],TestReciprocal->[setUp->[init_dtype]],TestSoftsign->[setUp->[init_dtype]],TestBRelu->[setUp->[init_dtype]],TestSqrt->[setUp->[init_dtype]],TestAbs->[setUp->[init_dtype]],TestFloor->[setUp->[init_dtype]],TestRelu->[setUp->[init_dtype]],TestPow->[setUp->[init_dtype]],TestSwish->[setUp->[init_dtype]],TestTanh->[setUp->[init_dtype]],TestLog->[setUp->[init_dtype]],TestSigmoid->[setUp->[init_dtype]],TestELU->[setUp->[init_dtype]],TestSoftShrink->[setUp->[init_dtype]],TestCos->[setUp->[init_dtype]],TestSTanh->[setUp->[init_dtype]],TestSquare->[setUp->[init_dtype]],TestSoftRelu->[setUp->[init_dtype]],TestSoftplus->[setUp->[init_dtype]],TestTanhShrink->[setUp->[init_dtype]],TestCeil->[setUp->[init_dtype]],TestRelu6->[setUp->[init_dtype]],TestRound->[setUp->[init_dtype]]] | Check if the object has the correct output. | Something to consider in the future returning None or Nan might be preferred since it avoids constructing a large zero tensor, saving some time and space. Just curious, why it works before this PR? |
@@ -87,7 +87,7 @@
public string EndpointName { get; }
- public IConfigureComponents Container { get; }
+ public IServiceCollection Container { get; }
public string DiagnosticsPath { get; }
| [HostingComponent->[Configuration->[IsINeedToInstallSomething]]] | AddStartupDiagnosticsSection - Add a section to the diagnostic section. | Should we rename `Container` to `Services`, it's not really a container anymore. |
@@ -258,7 +258,7 @@ Sagas must have at least one message that is allowed to start the saga. Add at l
throw new Exception(error);
}
- var existingMapping = mapper.Mappings.SingleOrDefault(m => m.MessageType == messageType);
+ var existingMapping = mapper.Mappings.SingleOrDefault(m => m.MessageType.IsAssignableFrom(m.MessageType));
if (existingMapping != null)
{
var bothMappingAndFinder = $"A custom IFindSagas and an existing mapping where found for message '{messageType.FullName}'. Either remove the message mapping for remove the finder. Finder name '{finderType.FullName}'.";
| [SagaMetadata->[ApplyScannedFinders->[MessageType,ConfigureCustomFinder,SingleOrDefault,IsMessageType,GetGenericArguments,GetInterfaces,ToList,FullName,Length],GetAssociatedMessages->[MessageType,Any,ToList,GetMessagesCorrespondingToFilterOnSaga,Add],TryGetFinder->[TryGetValue],SagaMapper->[ValidateMapping->[Format,PropertyType,Operand,NodeType,Name,FieldType,Convert,Body,Member],ThrowIfNotPropertyLambdaExpression->[Body],ConfigureCustomFinder->[Add],ConfigureMapping->[PropertyType,GetMemberInfo,Name,Compile,ThrowIfNotPropertyLambdaExpression,ValidateMapping,Add,compiledMessageExpression]],SetFinderForMessage->[MessageType,MakeGenericType,MessageProp,SagaPropName,HasCustomFinderMap,Add,CustomFinderType],IsMessageAllowedToStartTheSaga->[IsAllowedToStartSaga,TryGetValue],IsSagaType->[IsGenericType,IsAbstract,IsAssignableFrom],Type->[BaseType],GetMessagesCorrespondingToFilterOnSaga->[MakeGenericType,GetGenericArguments,GetInterfaces],correlationProperty,Any,Join,IsAllowedToStartSaga,Contains,Name,MessageTypeName,ToList,Select,Type]] | Applies the scanned Finders to the sagaMapper. | this exception mapping requires some improvements. where -> were. for -> or. |
@@ -18,4 +18,12 @@ class TrustLevelAndStaffSetting < TrustLevelSetting
def self.special_groups
['staff', 'admin']
end
+
+ def self.translation(value)
+ if special_group?(value)
+ I18n.t("trust_levels.#{value}")
+ else
+ super
+ end
+ end
end
| [TrustLevelAndStaffSetting->[valid_values->[to_a],valid_value?->[to_i,to_s,any?,special_group?],special_group?->[to_s,include?]]] | Returns a object for the special groups. | I think a test here will help prevent this from regressing in the future. |
@@ -79,7 +79,7 @@ class JsonParser(Parser):
This includes `namedtuple` subtypes as well as any custom class with an `_asdict` method defined;
see :class:`pants.engine.serializable.Serializable`.
"""
- json = filecontent
+ json = ensure_text(filecontent)
decoder = self._decoder
| [JsonParser->[_as_type->[_import],parse->[format_line,non_comment_line]]] | Parse the given json encoded string into a list of top - level objects found. Parse the JSON data file and return a list of objects. | I spent a lot of time on this part and am not 100% convinced I chose the best answer, although this works. `filecontent` was always unicode, except for a couple times it would equal `b''`, which I could never find the source for. So `encode()` and `decode()` don't work, since it's mixed. In general, I think we wanted `filecontent` to be binary type, but then with this being a JSON parser unicode makes sense and almost all our use cases had it as unicode. Either way, I think this file is only used in testing, right, because it falls under test/examples? |
@@ -174,7 +174,7 @@ public class HttpResponseBuilder extends HttpMessageBuilder implements Initialis
try
{
ByteArrayHttpEntity byteArrayHttpEntity = new ByteArrayHttpEntity(event.getMessage().getPayloadAsBytes());
- if (responseStreaming == ALWAYS || (responseStreaming == AUTO && CHUNKED.equals(existingTransferEncoding)))
+ if (responseStreaming == ALWAYS || (responseStreaming == AUTO && existingContentLength == null && CHUNKED.equals(existingTransferEncoding)))
{
setupChunkedEncoding(httpResponseHeaderBuilder);
}
| [HttpResponseBuilder->[initialise->[initialise],emptyInstance->[HttpResponseBuilder,init],build->[build]]] | Build the response. This method creates a new HttpEntity object based on the payload. Build the result. | Shouldn't we log in the case when both explicit headers are set. |
@@ -39,6 +39,7 @@ import {getServicePromise} from './service';
const SERVICES = {
'amp-dynamic-css-classes': '[custom-element=amp-dynamic-css-classes]',
'variant': 'amp-experiment',
+ 'amp-story-render': 'amp-story[standalone]',
};
/**
| [No CFG could be retrieved] | Provides a base class for rendering delaying services. Determines if a window is finished delaying render and is ready. | The validator team will need to update our list of render-blocking extensions so that this one is injected with a link-preload when necessary. |
@@ -417,7 +417,6 @@ pool_namecheck(const char *pool, namecheck_err_t *why, char *what)
return (0);
}
-#if defined(_KERNEL)
EXPORT_SYMBOL(entity_namecheck);
EXPORT_SYMBOL(pool_namecheck);
EXPORT_SYMBOL(dataset_namecheck);
| [dataset_nestcheck->[get_dataset_depth],dataset_namecheck->[entity_namecheck],permset_namecheck->[zfs_component_namecheck]] | This function checks if the name of a node is not too long and if so it can Returns a list of all the possible values for the current request. | do not remove kernel guard where we can use shared files on kernel side builds and in userland. for example (on DilOS): we are building zfs module files in userland with libzpool and should avoid kernel specific changes. |
@@ -485,10 +485,13 @@ class DepsCppInfo(_BaseDepsCppInfo):
def __init__(self):
super(DepsCppInfo, self).__init__()
self._dependencies = OrderedDict()
- self.configs = {}
+ self._configs = {}
def __getattr__(self, config):
- return self.configs.setdefault(config, _BaseDepsCppInfo())
+ return self._configs.setdefault(config, _BaseDepsCppInfo())
+
+ def get_configs(self):
+ return self._configs
@property
def dependencies(self):
| [CppInfo->[__getattr__->[_get_cpp_info->[_CppInfo],_get_cpp_info],__init__->[Component,DefaultOrderedDict]],DepCppInfo->[defines->[_aggregated_values],__getattr__->[_get_cpp_info->[],__getattr__],_get_sorted_components->[_filter_component_requires],cflags->[_aggregated_values],system_libs->[_aggregated_values],libs->[_aggregated_values],include_paths->[_aggregated_paths],res_paths->[_aggregated_paths],exelinkflags->[_aggregated_values],frameworks->[_aggregated_values],build_modules_paths->[_aggregated_paths],build_paths->[_aggregated_paths],src_paths->[_aggregated_paths],_aggregated_values->[_merge_lists],_aggregated_paths->[_merge_lists],lib_paths->[_aggregated_paths],framework_paths->[_aggregated_paths],bin_paths->[_aggregated_paths],sharedlinkflags->[_aggregated_values],cxxflags->[_aggregated_values],_check_component_requires->[_filter_component_requires]],_CppInfo->[lib_paths->[_filter_paths],framework_paths->[_filter_paths],include_paths->[_filter_paths],bin_paths->[_filter_paths],res_paths->[_filter_paths],build_paths->[_filter_paths],src_paths->[_filter_paths]],_BaseDepsCppInfo->[update->[merge_lists]],DepsCppInfo->[__getattr__->[_get_cpp_info->[],_BaseDepsCppInfo]],DefaultOrderedDict->[__copy__->[DefaultOrderedDict]]] | Initialize the dependencies CppInfo object. | I am curious why this is not just a property? |
@@ -58,11 +58,11 @@ class StudentsController < ApplicationController
def filter
case params[:filter]
when 'hidden'
- @students = Student.all(:conditions => {:hidden => true}, :order => :user_name)
+ @students = Student.all(conditions: {hidden: true}, order: :user_name)
when 'visible'
- @students = Student.all(:conditions => {:hidden => false}, :order => :user_name)
+ @students = Student.all(conditions: {hidden: false}, order: :user_name)
else
- @students = Student.all(:order => :user_name)
+ @students = Student.all(order: :user_name)
end
end
| [StudentsController->[upload_student_list->[redirect_to,size,upload_user_list,post?,blank?,t],create->[redirect_to,save,post?,new],edit->[find_by_id],update->[redirect_to,render,user_name,update_attributes,post?,find_by_id],bulk_modify->[hide_students,nil?,empty?,find,give_grace_credits,construct_table_rows,render,unhide_students,message,raise],filter->[all],download_student_list->[generate_csv_list,to_xml,all,send_data],populate->[construct_table_rows,all],index->[all],include,before_filter]] | This action is used to filter the user by sequence number and optionally by the students. | Use 2 (not 3) spaces for indentation.<br>Space inside { missing.<br>Space inside } missing. |
@@ -50,7 +50,11 @@ export default class SideNav extends React.Component<NavPropsT> {
StyledSubNavContainer,
);
- const renderNavItem = (item: Item, level: number, index) => {
+ const renderNavItem = (item: Item, level: number, index, transformItem) => {
+ if (typeof transformItem === 'function') {
+ item = transformItem(item);
+ }
+
const sharedProps = {
$active: activePredicate
? activePredicate(item, activeItemId)
| [No CFG could be retrieved] | A component that renders a side - nav. Create a that represents the nav items of the given item. | @chasestarr This will run on each render which might not be what the user wants. For example it will add the prefix again on each render. ~~Also Is see that the component doesn't re-render if the items array is changed, is that intended?~~ it does :) |
@@ -77,3 +77,15 @@ class Image(graphene.ObjectType):
url = image.url
url = info.context.build_absolute_uri(url)
return Image(url, alt)
+
+
+class PriceInput(graphene.InputObjectType):
+ gte = graphene.Float(
+ description='Price greater than or equal', required=False)
+ lte = graphene.Float(
+ description='Price less than or equal', required=False)
+
+
+class DateRangeInput(graphene.InputObjectType):
+ from_date = graphene.Date(description='Starting date from', required=False)
+ to_date = graphene.Date(description='To date', required=False)
| [PermissionDisplay->[PermissionEnum,String],Image->[get_adjusted->[build_absolute_uri,get_thumbnail,Image],String],Error->[String,dedent],LanguageDisplay->[LanguageCodeEnum,String],SeoInput->[String],Weight->[String,Float],CountryDisplay->[String,Field]] | Return Image adjusted with given size. | If we call below `DateRangeInput`, maybe this one could be `PriceRangeInout`? Or the previous one just `DateInput`? Would be more consistent IMO. |
@@ -62,13 +62,7 @@
current_position.set(0.0, 0.0);
sync_plan_position();
- const int x_axis_home_dir =
- #if ENABLED(DUAL_X_CARRIAGE)
- x_home_dir(active_extruder)
- #else
- home_dir(X_AXIS)
- #endif
- ;
+ const int x_axis_home_dir = x_home_dir(active_extruder);
const float mlx = max_length(X_AXIS),
mly = max_length(Y_AXIS),
| [No CFG could be retrieved] | Quick home of the system. region ethernet interface. | Does this line behave like previous? Can't analyze function contents now |
@@ -91,9 +91,8 @@ func (s *REST) Create(ctx kapi.Context, obj runtime.Object) (runtime.Object, err
processor := template.NewProcessor(generators)
cfg, err := processor.Process(tpl)
if len(err) > 0 {
- // TODO: We don't report the processing errors to users as there is no
- // good way how to do it for just some items.
glog.V(1).Infof(utilerr.NewAggregate(err).Error())
+ return nil, errors.NewInvalid("template", tpl.Name, err)
}
if tpl.ObjectLabels != nil {
| [Validate->[ValidateTemplate],ValidateUpdate->[ValidateTemplateUpdate],Create->[NewInvalid,NewProcessor,UnixNano,NewSource,Infof,Error,NewAggregate,AddConfigLabels,New,Now,ValidateProcessedTemplate,V,Process,NewBadRequest,Set,NewExpressionValueGenerator],MatcherFunc,Set,Matches,Errorf] | Create creates a new template object. | Remove this - we don't need it now |
@@ -64,7 +64,7 @@ func (s *policyServer) migrateV1Policies(ctx context.Context) ([]error, error) {
if storagePol == nil {
continue // nothing to create
}
- _, err = s.store.CreatePolicy(ctx, storagePol, false)
+ _, err = s.store.CreatePolicy(ctx, storagePol, true)
switch err {
case nil, storage_errors.ErrConflict: // ignore, continue
default:
| [addTokenToAdminPolicy->[Wrap,Wrapf,AddPolicyMembers,NewMember],okToMigrate->[Info,Error],migrateV1Policies->[Wrapf,Error,CreatePolicy,String,Errorf,addTokenToAdminPolicy,ListPoliciesWithSubjects],NewPolicy,Wrapf,Join,Sprintf,Split,NewStatement,New,NewMember,String,DefaultPolicies,Errorf,Wrap,HasPrefix] | migrateV1Policies migrates all v1 policies in the store. | This surprised me at first... but then I saw that you reversed polarity of that boolean. |
@@ -30,7 +30,7 @@ class XmlParser:
try:
parsed_xml = cls._parse(xml_path)
except OSError as e:
- raise XmlParser.XmlError("Problem reading xml file at {}: {}".format(xml_path, e))
+ raise XmlParser.XmlError(f"Problem reading xml file at {xml_path}: {e}")
return cls(xml_path, parsed_xml)
def __init__(self, xml_path, parsed_xml):
| [XmlParser->[from_file->[XmlError,_parse],get_attribute->[XmlError],get_optional_attribute->[get_attribute],_parse->[XmlError]]] | Parse. xml file and create a XmlParser object. | In general, great to use `{e!r}` when printing exceptions because it will ensure that the exception class is in the message. We don't do a great job at doing this, but some Pycon talk on exceptions mentioned it's a best practice. |
@@ -32,9 +32,9 @@ namespace DynamoWebServer
webSocketServer.Start();
- var app = new Application();
- app.Exit += webSocketServer.ProcessExit;
- app.Run();
+ Process.GetCurrentProcess().Exited += webSocketServer.ProcessExited;
+
+ while (true) {}
}
}
}
| [Program->[Main->[Exit,MaxTesselationDivisions,PreloadAsmLibraries,AppSettings,ProcessExit,Start,Location,Parse,Run,GetDirectoryName,Instance,Load,InitializeCore]]] | Main method that runs the application. | I'm sure this is temporary and that we don't have a terminating condition, do we want to do this for the server and enhance this in the next PR? |
@@ -1203,8 +1203,8 @@ namespace ProtoCore.Lang
return DSASM.StackValue.Null;
List<StackValue> svList = new List<StackValue>();
- StackValue[] svArray1 = runtime.runtime.rmem.GetArrayElements(sv1);
- StackValue[] svArray2 = runtime.runtime.rmem.GetArrayElements(sv2);
+ var svArray1 = ArrayUtils.GetValues(sv1, runtime.runtime.Core);
+ var svArray2 = ArrayUtils.GetValues(sv2, runtime.runtime.Core);
svList.AddRange(svArray1);
svList.AddRange(svArray2);
| [StackValueComparerForDouble->[Compare->[Equals],Equals->[Equals]],RangeExpressionUntils->[StackValue->[GenerateRangeByStepNumber,GenerateRangeByAmount,GenerateRangeByStepSize]],FileIOBuiltIns->[StackValue->[ConvertToString]],ArrayUtilsForBuiltIns->[CountTrue->[CountTrue],IsUniformDepth->[Rank],Exists->[Exists],EqualsInValue->[EqualsInValue],SomeFalse->[Exists],SomeNulls->[Exists],Rank->[Rank],SomeTrue->[Exists],ForAll->[ForAll],GetFlattenedArrayElements->[GetFlattenedArrayElements],IsHomogeneous->[IsHomogeneous],Contains->[EqualsInValue,Contains],AllTrue->[ForAll],IsRectangular->[Rank,Count],ContainsArray->[EqualsInValue,ContainsArray],IndexOf->[EqualsInValue],ArrayIndexOfArray->[EqualsInValue],CountFalse->[CountFalse],StackValue->[ConvertToString,Rank,Count,CountNumber],AllFalse->[ForAll]]] | Concat method for DSASM. StackValue. Concat. | You can use svArray1.Concat(svArray2).ToArray() to avoid svList.ToArray(); |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.