patch
stringlengths
18
160k
callgraph
stringlengths
4
179k
summary
stringlengths
4
947
msg
stringlengths
6
3.42k
@@ -92,6 +92,18 @@ public final class ClassLoaderMatcher { } } + private static boolean canSkipClassLoaderByPackagePrefix(ClassLoader loader) { + String name = loader.getClass().getName(); + if (name.startsWith("datadog.") + || name.startsWith("com.dynatrace.") + || name.startsWith("com.appdynamics.") + || name.startsWith("com.newrelic.") + || name.startsWith("com.nr.agent.")) { + return true; + } + return false; + } + /** * TODO: this turns out to be useless with OSGi: {@code * org.eclipse.osgi.internal.loader.BundleLoader#isRequestFromVM} returns {@code true} when
[ClassLoaderMatcher->[SkipClassLoaderMatcher->[SkipClassLoaderMatcher],ClassLoaderHasClassesNamedMatcher->[matches->[hasResources]]]]
Checks if the given class loader can be skipped by this class loader.
I would just add this into `canSkipClassLoaderByName`
@@ -631,8 +631,16 @@ def _auto_topomap_coords(info, picks): locs3d = np.array([eeg_ch_locs[ch['ch_name']] for ch in chs]) # Duplicate points cause all kinds of trouble during visualization - if np.min(pdist(locs3d)) < 1e-10: - raise ValueError('Electrode positions must be unique.') + dist = pdist(locs3d) + if np.min(dist) < 1e-10: + problematic_electrodes = [ + info['ch_names'][elec_i] + for elec_i in np.unique(squareform(dist < 1e-10).nonzero()[0]) + ] + + raise ValueError('The following electrodes have overlapping positions:' + '\n ' + str(problematic_electrodes) + '\nThis ' + 'causes problems during visualization.') x, y, z = locs3d.T az, el, r = _cartesian_to_sphere(x, y, z)
[make_grid_layout->[Layout],_box_size->[ydiff,xdiff],find_layout->[read_layout,make_eeg_layout],read_layout->[_read_lout,Layout,_read_lay],make_eeg_layout->[Layout],_pair_grad_sensors->[_find_topomap_coords],generate_2d_layout->[Layout,_box_size]]
Auto - generate a 2 dimensional sensor map from sensor positions in an info dict. Return the 2 - D array of possible digitization points that are in the 3 - D.
I suspect `np.unique(squareform(dist < 1e-10))` will be slower than `(dist < 1e-10).any(axis=0)`, and a bit easier to understand, assuming I'm right in thinking they're equivalent. WDYT?
@@ -90,6 +90,8 @@ public abstract class KafkaSource<S, D> extends EventBasedSource<S, D> { public static final String LATEST_OFFSET = "latest"; public static final String EARLIEST_OFFSET = "earliest"; public static final String NEAREST_OFFSET = "nearest"; + public static final String OFFSET_LOOKBACK = "offset_lookback"; + public static final String TIMESTAMP_LOOKBACK = "timestamp_lookback"; public static final String BOOTSTRAP_WITH_OFFSET = "bootstrap.with.offset"; public static final String DEFAULT_BOOTSTRAP_WITH_OFFSET = LATEST_OFFSET; public static final String TOPICS_MOVE_TO_LATEST_OFFSET = "topics.move.to.latest.offset";
[KafkaSource->[getWorkUnitForTopicPartition->[addDatasetUrnOptionally,getWorkUnitForTopicPartition],getFilteredTopics->[getFilteredTopics],getWorkunits->[getLimiterExtractorReportKeys,setLimiterReportKeyListToWorkUnits],addTopicSpecificPropsToWorkUnit->[addTopicSpecificPropsToWorkUnit],WorkUnitCreator->[run->[getWorkUnitsForTopic]],createEmptyWorkUnit->[getWorkUnitForTopicPartition]]]
Implementation for Kafka source. region Time - String - Property names.
seems like timestamp lookback is not being used. Do you plan to modify this PR with support for timestamp lookback as well?
@@ -24,10 +24,12 @@ import sys import textwrap import time import traceback - import six +import types +import llnl.util.filesystem as fsys import llnl.util.tty as tty + import spack.compilers import spack.config import spack.dependency
[PackageBase->[do_patch->[do_stage],_if_make_target_execute->[_has_make_target],setup_build_environment->[_get_legacy_environment_method],do_stage->[do_fetch],dependency_activations->[extends],do_deactivate->[is_activated,_sanity_check_extension,do_deactivate],do_uninstall->[uninstall_by_spec],possible_dependencies->[possible_dependencies],_if_ninja_target_execute->[_has_ninja_target],setup_run_environment->[_get_legacy_environment_method],do_deprecate->[do_deprecate,uninstall_by_spec,do_install,is_activated],do_activate->[_sanity_check_extension,is_activated,do_activate],fetcher->[_make_fetcher],nearest_url->[version_urls],_make_stage->[_make_resource_stage,_make_root_stage],setup_dependent_run_environment->[_get_legacy_environment_method],_sanity_check_extension->[_check_extendable],stage->[_make_stage],setup_dependent_build_environment->[_get_legacy_environment_method],sanity_check_prefix->[check_paths],url_for_version->[nearest_url,version_urls],on_package_attributes],InstallPhase->[copy->[InstallPhase]],possible_dependencies->[possible_dependencies],run_before->[register_callback],run_after->[register_callback],PackageMeta->[__new__->[_flush_callbacks->[copy],_flush_callbacks,InstallPhase]],Package->[run_after]]
Copyright 2013 - 2013 - 2013 - 2013 - httpd URL - Schemes for the Spack build.
This is always imported `as fs`
@@ -3,9 +3,14 @@ package utils import ( "bytes" "encoding/binary" + "encoding/json" + "fmt" + "io" "log" + "os" "regexp" "strconv" + "sync" "github.com/dedis/kyber" "github.com/harmony-one/harmony/crypto"
[Write,ReplaceAllString,SetInt64,Panic,Scalar,Compile,GetPublicKeyFromScalar,Bytes,Atoi]
ConvertFixedDataIntoByteArray converts an empty interface to a byte array. AllocateShard returns the number of current nodes and number of shards.
Why not `func Unmarshal(r io.Reader, v interface{}) error { }`? Also, do you have any plan to reuse this function from outside of the `internal/utils` package? I wonder if it could be made private.
@@ -65,3 +65,8 @@ class Jsoncpp(CMakePackage): else: args.append('-DJSONCPP_WITH_TESTS=OFF') return args + + def check(self): + # TODO: known test failures + # see https://github.com/spack/spack/issues/24773 + pass
[Jsoncpp->[cmake_args->[append,format],patch->[filter_file],when,variant,depends_on,version]]
Return a list of arguments to pass to cmake.
Hmm, the issue says that the tests are failing - not why they are failing. If the failure is legitimate and not an issue known to upstream, we shouldn't skip this check since it's not a false positive.
@@ -113,6 +113,7 @@ class TestWikiTablesDatasetReader(AllenNlpTestCase): "<e,r> -> fb:row.row.avg_attendance", "<e,r> -> fb:row.row.division", "<e,r> -> fb:row.row.league", + "<e,r> -> fb:row.row.null", # null row, representing an empty set "<e,r> -> fb:row.row.open_cup", "<e,r> -> fb:row.row.playoffs", "<e,r> -> fb:row.row.regular_season",
[TestWikiTablesDatasetReader->[test_reader_reads->[len,sorted,read,isinstance,fields,WikiTablesDatasetReader]]]
Test the reading of the Wikitables dataset reads for a single . Order of time - sequence sequence. "d. type - > d. type - > d. type - > d. END of grammar grammar We don t need to do anything if we are in our list of valid actions.
You mean a "null column"? Also, why is this produced?
@@ -663,7 +663,7 @@ class Stage(object): self._check_dvc_filename(fname) - logger.info( + logger.debug( "Saving information to '{file}'.".format(file=relpath(fname)) ) d = self.dumpd()
[Stage->[_run->[_warn_if_fish,_check_missing_deps,StageCmdFailedError],dumpd->[relpath],_check_stage_path->[StagePathOutsideError,StagePathNotFoundError,StagePathNotDirectoryError],_status->[update],_changed_md5->[changed_md5],relpath->[relpath],is_stage_file->[is_valid_filename],remove->[unprotect_outs,remove_outs],_check_file_exists->[StageFileDoesNotExistError],check_missing_outputs->[MissingDataSource],check_can_commit->[save,changed_md5,StageCommitError,_changed_entries],dump->[relpath,dumpd,_check_dvc_filename],update->[StageUpdateError,reproduce],run->[save,commit,_run,remove_outs],status->[changed_md5,_status],is_cached->[_changed_outs],load->[validate,relpath,Stage,_check_file_exists,_check_dvc_filename,_check_isfile,_get_path_tag],_compute_md5->[dumpd],save->[save,_compute_md5],_check_missing_deps->[MissingDep],_check_dvc_filename->[relpath,is_valid_filename,StageFileBadNameError],checkout->[checkout],_check_isfile->[StageFileIsNotDvcFileError],_already_cached->[changed_md5,changed],commit->[commit],create->[_stage_fname,_check_stage_path,StageFileBadNameError,relpath,Stage,StageFileAlreadyExistsError,remove_outs],changed->[_changed_deps,_changed_outs,_changed_md5],reproduce->[changed],validate->[StageFileFormatError]]]
Dumps all the data of the object to a file.
And one last question: Is this really a debug output? I also find that message useful in the output.
@@ -868,8 +868,10 @@ class JarPublish(TransitiveOptionRegistrar, HasTransitiveOptionMixin, ScmPublish def exportable(tgt): return tgt in candidates and self._is_exported(tgt) - return OrderedSet(filter(exportable, - reversed(sort_targets(filter(exportable, candidates))))) + return OrderedSet(target for target in + reversed(sort_targets(candidate for candidate in candidates + if exportable(candidate))) + if exportable(target)) def entry_fingerprint(self, target, fingerprint_internal): sha = hashlib.sha1()
[target_internal_dependencies->[target_internal_dependencies],jar_coordinate->[coordinate],pushdb_coordinate->[version,jar_coordinate],JarPublish->[generate_ivy->[write],changelog->[changelog],check_targets->[collect_invalid->[_is_exported],check_for_duplicate_artifacts],exported_targets->[exportable->[_is_exported],get_synthetic],check_for_duplicate_artifacts->[duplication_message,DuplicateArtifactError],create_doc_jar->[add_docs->[write],_scala_doc,artifact_path,add_docs,_java_doc],entry_fingerprint->[target_internal_dependencies,jar_coordinate,fingerprint_internal],create_source_jar->[write,artifact_path,abs_and_relative_sources],__init__->[parse_override->[parse_jarcoordinate],parse_jarcoordinate->[_is_exported],parse_override,parse_jarcoordinate],generate_ivysettings->[fetch_ivysettings,write],execute->[get_pushdb->[get_db],stage_artifacts->[write,_copy_artifact,Publication,PomWriter],fingerprint_internal->[get_pushdb,get_entry],get_db->[load,PushDb],confirm_push,get_entry,jar_coordinate,write,pushdb_coordinate,stage_artifacts,with_named_ver,coordinate,version,set_entry,with_sem_ver,with_sha_and_fingerprint,dump,publish,get_db],publish->[version,_ivy_jvm_options,pushdb_coordinate]],PushDb->[_accessors_for_target->[setter->[key],getter->[key]],load->[load,PushDb],get_entry->[Entry],set_entry->[version],Entry->[with_sem_ver->[Entry],with_sha_and_fingerprint->[Entry],with_named_ver->[Entry]],dump->[dump]],PomWriter->[write->[write],_as_versioned_jar->[version]]]
Returns a list of all targets that can be exported.
This nested generator clause works, although is a little hard to read, as was the original. I can refactor if you want, although might be better to keep original structure.
@@ -361,6 +361,11 @@ export class SystemLayer { */ onUnmuteAudioClick_() { this.storeService_.dispatch(Action.TOGGLE_MUTED, false); + const hideTimeout = 1500; + this.vsync_.mutate(() => { + this.getShadowRoot().setAttribute('messagedisplay', 'show'); + this.hideAfterTimeout_(hideTimeout); + }); } /**
[No CFG could be retrieved]
Provides a way to handle has audio state updates. This method updates the progress of the and creates the share pill if necessary.
This method looks duplicated
@@ -524,6 +524,14 @@ class Grouping < ApplicationRecord collection_date < Time.current end + # Return the duration of this grouping's assignment or the time between now and the collection_date whichever is + # less. If the collection date has already passed, return 0 seconds. + def adjusted_duration + add = assignment.submission_rule.periods.pluck(:hours).sum.hours + (extension&.time_delta || 0) + start = [assignment.section_start_time(inviter&.section), Time.current].max + assignment.assignment_properties.adjusted_duration(collection_date, start, add: add) + end + def self.get_assign_scans_grouping(assignment, grouping_id = nil) subquery = StudentMembership.all.to_sql assignment.groupings.includes(:non_rejected_student_memberships)
[Grouping->[deletable_by?->[is_valid?],test_runs_instructors->[pluck_test_runs,group_hash_list,filter_test_runs],has_files_in_submission?->[has_submission?],due_date->[due_date],test_runs_students_simple->[filter_test_runs],remove_member->[membership_status],remove_rejected->[membership_status],test_runs_students->[pluck_test_runs,group_hash_list,filter_test_runs],can_invite?->[pending?],membership_status->[membership_status],test_runs_instructors_released->[pluck_test_runs,group_hash_list,filter_test_runs]]]
Checks if the given collection date is past the last unknown groupings.
I'm thinking the submission rule periods shouldn't be counted as part of the reported duration, but rather listed separately.
@@ -375,13 +375,13 @@ static int dtrace_calc_buf_overflow(struct dma_trace_buf *buffer, uint32_t overflow_margin; uint32_t overflow = 0; - margin = buffer->end_addr - buffer->w_ptr; + margin = dtrace_calc_buf_margin(buffer); /* overflow calculating */ if (buffer->w_ptr < buffer->r_ptr) - overflow_margin = buffer->r_ptr - buffer->w_ptr; + overflow_margin = buffer->r_ptr - buffer->w_ptr - 1; else - overflow_margin = margin + buffer->r_ptr - buffer->addr; + overflow_margin = margin + buffer->r_ptr - buffer->addr - 1; if (overflow_margin < length) overflow = length - overflow_margin;
[dma_trace_init_complete->[work_init,trace_buffer,trace_buffer_error,dma_copy_new],int->[dcache_writeback_invalidate_region,ipc_dma_trace_send_position,dma_set_config,dma_start,trace_buffer_error,rballoc,dma_copy_set_stream_tag,dma_sg_alloc,bzero],dma_trace_enable->[dma_trace_buffer_init,trace_error_atomic,dma_trace_start,work_schedule_default],dtrace_event->[spin_unlock_irq,cpu_get_id,spin_lock_irq,work_reschedule_default,dtrace_add_event],dma_trace_flush->[memcpy,dcache_writeback_invalidate_region],dma_trace_init_early->[spinlock_init,dma_sg_init,rzalloc],dtrace_event_atomic->[dtrace_add_event],void->[trace_error,dcache_writeback_invalidate_region,dtrace_calc_buf_overflow,memcpy,dtrace_calc_buf_margin],uint64_t->[spin_unlock_irq,spin_lock_irq,dma_trace_get_avail_data,trace_buffer_error,dma_copy_to_host_nowait]]
Calculate the overflow of the buffer.
Sorry, I'm not following the commit message. What is the bug this is fixing and why does -1 fix it ? Thanks
@@ -208,9 +208,9 @@ func orgAssignment(args ...bool) macaron.Handler { var err error if assignOrg { - ctx.Org.Organization, err = models.GetUserByName(ctx.Params(":orgname")) + ctx.Org.Organization, err = models.GetOrgByName(ctx.Params(":orgname")) if err != nil { - if models.IsErrUserNotExist(err) { + if models.IsErrOrgNotExist(err) { ctx.Status(404) } else { ctx.Error(500, "GetUserByName", err)
[IsOrganizationOwner,EnableUnit,Patch,Status,APIContexter,Delete,RepoRef,IsWriter,GetRepositoryByName,AllowsPulls,IsErrRepoNotExist,Put,IsErrRepoRedirectNotExist,Error,GetTeamByID,Post,IsOrganizationMember,RedirectToRepo,LookupRepoRedirect,Any,ParamsInt64,ToLower,IsErrUserNotExist,GetUserByName,Get,HasAccess,ReferencesGitRepo,AccessLevel,Combo,Group,Params]
orgAssignment returns a middleware that assigns a user to an organization or team. RegisterRoutes registers all v1 routes to web application.
Please also fix 500 error name
@@ -759,7 +759,7 @@ d_tm_increment_counter(struct d_tm_node_t **metric, char *item, ...) } if (node == NULL) { - rc = d_tm_add_metric(&node, path, D_TM_COUNTER, "N/A", "N/A"); + rc = d_tm_add_metric(&node, protect, path, D_TM_COUNTER, "N/A", "N/A"); if (rc != D_TM_SUCCESS) { D_ERROR("Failed to add and incremement counter [%s]: " "rc = %d\n", path, rc);
[No CFG could be retrieved]
Adds a new metric or increases a counter. This function increments the current counter on the specified item.
(style) line over 80 characters
@@ -1,5 +1,5 @@ -<script> - function add_annotation_prompt(path) { +<%= javascript_tag nonce: true do %> +function add_annotation_prompt(path) { var annotation_text = $('#add_annotation_text #annotation_text_content'); if (annotation_text.length) {
[No CFG could be retrieved]
Adds a new annotation prompt to the menu.
This is small but I'd prefer leaving the indentation as-is
@@ -30,11 +30,17 @@ class Mesa(AutotoolsPackage): specification - a system for rendering interactive 3D graphics.""" homepage = "http://www.mesa3d.org" - url = "http://ftp.iij.ad.jp/pub/X11/x.org/pub/mesa/12.0.3/mesa-12.0.3.tar.gz" + #url = "http://ftp.iij.ad.jp/pub/X11/x.org/pub/mesa/12.0.3/mesa-12.0.3.tar.gz" + url = "ftp://ftp.freedesktop.org/pub/mesa/older-versions/10.x/10.2.4/MesaLib-10.2.4.tar.bz2" version('12.0.3', '60c5f9897ddc38b46f8144c7366e84ad') + version('10.2.4', '11d3542da1b703618634be840a87b0b2') # General dependencies + depends_on('autoconf', type='build') + depends_on('automake', type='build') + depends_on('libtool', type='build') + depends_on('m4', type='build') depends_on('python@2.6.4:') depends_on('py-mako@0.3.4:') depends_on('flex@2.5.35:', type='build')
[Mesa->[depends_on,version]]
Creates a new object with the details of the given object. Package manager for the package manager.
why do you change `url`, is `10.2.4` not available from `iij.ad.jp`? If so, just add it to the `version()` and keep the old `url`.
@@ -75,6 +75,13 @@ module GobiertoBudgets mean_province_query(@year, "amount") end + def execution_percentage(requested_year = year) + @execution_percentage ||= begin + variable_2 = amount_updated ? :amount_updated : :amount_planned + percentage_difference(variable1: :amount_executed, variable2: variable_2, year: requested_year) + end + end + private def total_budget
[BudgetLineStats->[percentage_of_total->[amount]]]
Returns the mean province nexus query for this year.
Use normalcase for variable numbers.
@@ -113,6 +113,13 @@ public class SourceColumnManager implements CommandPaletteEntrySource, return JsUtil.toStringArray(getNamesNative()); } + public final native String getActiveColumn() /*-{ + if (this.activeColumn) + return this.activeColumn; + else + return ""; + }-*/; + private native JsArrayString getNamesNative() /*-{ return this.names; }-*/;
[SourceColumnManager->[doActivateSource->[hasActiveEditor,execute],saveAllSourceDocs->[execute,getValue],onPreviousTab->[getValue],isEmpty->[getTabCount],switchToTab->[getTabCount,getPhysicalTabIndex],saveEditingTargetsWithPrompt->[showUnsavedChangesDialog,execute],onSwitchToTab->[getTabCount,showOverflowPopout],editFile->[onResponseReceived->[onSuccess->[],openFile]],fireDocTabsChanged->[fireDocTabsChanged],consolidateColumns->[onResponseReceived->[onSuccess->[],addTab],createState,getNames],saveActiveSourceDoc->[hasActiveEditor],openFile->[execute->[onError->[],onResponseReceived->[onSuccess->[]],onCancelled],newDoc,onCancelled,openFile,ensureVisible,add],onMoveTabRight->[getPhysicalTabIndex],newRMarkdownV2Doc->[execute->[onError->[],onResponseReceived->[onSuccess->[]],newRMarkdownV2Doc]],withTarget->[withTarget,execute,findEditor],activateObjectExplorer->[onResponseReceived->[onSuccess->[],addTab]],getUntitledNum->[getUntitledNum],recordCurrentNavigationHistoryPosition->[hasActiveEditor],onInit->[createState,getNames],getEditorContext->[getEditorContext],reindent->[hasActiveEditor,reindent],manageCommands->[manageCommands],onSourceExtendedTypeDetected->[findEditor],onFirstTab->[getTabCount],attemptTextEditorActivate->[hasActiveEditor],reflowText->[reflowText,hasActiveEditor],revertActiveDocument->[hasActiveEditor],saveChanges->[findEditor,add],closeSourceDoc->[closeTab,getTabCount],isActiveEditor->[hasActiveEditor],onMoveTabToLast->[getTabCount,getPhysicalTabIndex],ensureVisible->[ensureVisible,getActive],showOverflowPopout->[showOverflowPopout],closeTab->[closeTab],showDataItem->[onResponseReceived->[onSuccess->[],addTab]],inEditorForId->[execute,findEditor],getPhysicalTabIndex->[getPhysicalTabIndex,getActive],closeAllTabs->[execute->[onError->[],onResponseReceived->[onSuccess->[]],closeTab,execute],closeAllTabs],onMoveTabLeft->[getPhysicalTabIndex],onDebugModeChanged->[hasActiveEditor],getWidgets->[add],showUnsavedChangesDialog->[showUnsavedChangesDialog],onNextTab->[getValue],closeAllLocalSourceDocs->[closeAllTabs],openFileFromServer->[onError->[onFailure,execute],onResponseReceived->[onSuccess->[],onSuccess,execute,addTab,add],getValue],openProjectDocs->[onExecute->[execute->[onError->[],onResponseReceived->[onSuccess->[]],execute],openFile,onFirstTab,execute],setActive],disownDoc->[findByDocument],disownDocOnDrag->[getSize,getActive],addTab->[addTab],processOpenFileQueue->[onFailure->[execute],onSuccess->[execute],execute->[onError->[],onResponseReceived->[onSuccess->[]],isEmpty,processOpenFileQueue],onCancelled->[execute,onCancelled],openFile,isEmpty],vimSetTabIndex->[getTabCount],onMoveTabToFirst->[getPhysicalTabIndex],activateCodeBrowser->[onSuccess->[onSuccess],onFailure->[onFailure],onCancelled->[onCancelled],hasActiveEditor,onSuccess],getEditorPositionString->[hasActiveEditor],newSourceDocWithTemplate->[onError->[onError],onResponseReceived->[onSuccess->[execute],transform],newSourceDocWithTemplate],showHelpAtCursor->[hasActiveEditor,showHelpAtCursor],add->[createState,getNames,add],getTabCount->[getTabCount],cpsExecuteForEachEditor->[onExecute->[execute->[onError->[],onResponseReceived->[onSuccess->[]]],execute],cpsExecuteForEachEditor],saveAndCloseActiveSourceDoc->[execute->[onError->[],onResponseReceived->[onSuccess->[]],onCloseSourceDoc],hasActiveEditor],onFindInFiles->[hasActiveEditor],setActive->[setActive],getNames->[add],getCommandPaletteItems->[hasActiveEditor,getCommandPaletteItems],activateColumns->[onSuccess->[setActive],setActive],initVimCommands->[closeAllTabs,getValue],revertUnsavedTargets->[closeTab],newDoc->[newDoc],initialSelect->[initialSelect],setActiveDocId->[setActive],inEditorForPath->[execute,findEditorByPath],insertSource->[hasActiveEditor],openFileAlreadyOpen->[onSuccess,selectTab,add],onLastTab->[getTabCount],closeColumn->[createState,getTabCount,setActive,getNames],hasDoc->[hasDoc],openNotebook->[execute->[onError->[onFailure],onResponseReceived->[onSuccess->[],openNotebook]],onError->[onFailure],onResponseReceived->[onSuccess->[],onSuccess,addTab]],closeTabWithPath->[findEditorByPath],pasteRCodeExecutionResult->[onResponseReceived->[onSuccess->[],hasActiveEditor]],closeTabs->[closeTabs],selectTab->[selectTab,findByDocument],onSwitchToDoc->[execute],getActive->[setActive],initDynamicCommands->[add]]]
Returns the names of the node in the order they were specified.
IMHO, a more idiomatic way of specifying this is JavaScript would be `this.activeColumn || ""`.
@@ -84,6 +84,8 @@ co_create(void **state) /** destroy container */ if (arg->myrank == 0) { + /* XXX check if this is a real leak or out-of-sync close */ + sleep(5); print_message("destroying container %ssynchronously ...\n", arg->async ? "a" : ""); rc = daos_cont_destroy(arg->pool.poh, uuid, 1 /* force */,
[void->[handle_share,uuid_generate,daos_cont_list_attr,strdup,daos_event_init,MPI_Barrier,daos_prop_free,test_teardown,daos_cont_get_attr,daos_cont_destroy,daos_cont_close,test_setup_next_step,WAIT_ON_ASYNC,daos_cont_query,strlen,daos_mgmt_set_params,daos_event_fini,daos_pool_query,assert_int_equal,daos_prop_alloc,test_setup,print_message,strcmp,daos_cont_create,daos_cont_set_attr,assert_memory_equal,assert_true,daos_prop_entry_get,assert_string_equal,daos_cont_open,ARRAY_SIZE],int->[test_setup,async_disable,async_enable],run_daos_cont_test->[MPI_Barrier,cmocka_run_group_tests_name]]
private static final int REAL_CO = 0 ; destroy a container asynchronously.
probably can remove this. BTW, for original issue adding a sleep before cont_close maybe helpful, now not needed.
@@ -17,7 +17,12 @@ package notests -import "fmt" +import ( + "fmt" + // mage:import + + integtest "github.com/elastic/beats/v7/dev-tools/mage/target/integtest" +) // IntegTest executes integration tests (it uses Docker to run the tests). func IntegTest() {
[Println]
IntogTest executes the integration tests of a single .
I think this line is not needed here.
@@ -24,8 +24,8 @@ class ArchiveLinkResource(Resource): resource_title = endpoint_name schema = { - 'link_id': Resource.rel('archive', False, required=False), - 'desk': Resource.rel('desks', required=False) + 'link_id': Resource.rel('archive', False, type='string'), + 'desk': Resource.rel('desks', False) } url = 'archive/<{0}:target_id>/link'.format(item_url)
[ArchiveLinkService->[create->[get_resource_service,link_as_next_take,get,update,find_one],TakesPackageService],ArchiveLinkResource->[rel,format],getLogger]
Create a link to an existing object in the archive.
It might be a good idea to retain the `required` argument keyword so that's it's immediately clear what that `False` argument stands for.
@@ -193,6 +193,18 @@ func (cs *ContainerService) setKubeletConfig(isUpgrade bool) { } } + if isUpgrade && common.IsKubernetesVersionGe(o.OrchestratorVersion, "1.14.0") { + hasSupportPodPidsLimitFeatureGate := strings.Contains(o.KubernetesConfig.KubeletConfig["--feature-gates"], "SupportPodPidsLimit=true") + podMaxPids, _ := strconv.Atoi(profile.KubernetesConfig.KubeletConfig["--pod-max-pids"]) + if podMaxPids > 0 { + // If we don't have an explicit SupportPodPidsLimit=true, disable --pod-max-pids by setting to -1 + // To prevent older clusters from inheriting SupportPodPidsLimit=true implicitly starting w/ 1.14.0 + if !hasSupportPodPidsLimitFeatureGate { + profile.KubernetesConfig.KubeletConfig["--pod-max-pids"] = strconv.Itoa(-1) + } + } + } + removeKubeletFlags(profile.KubernetesConfig.KubeletConfig, o.OrchestratorVersion) } }
[setKubeletConfig->[Itoa,IsIPMasqAgentEnabled,IsKubernetesVersionGe,Contains,IsNVIDIADevicePluginEnabled,Bool,Atoi],IsKubernetesVersionGe]
setKubeletConfig sets the kubelet config This function is used to configure the static Windows Kubelet. This function is used to configure the kubelet with the given options. VersionGe sets up the necessary config values for the kubelet based on the version of the or 1.
This should get the --feature-gates flag from the agent profile
@@ -2279,7 +2279,7 @@ migrate_cont_iter_cb(daos_handle_t ih, d_iov_t *key_iov, arg.pool_tls = tls; uuid_copy(arg.cont_uuid, cont_uuid); while (!dbtree_is_empty(root->root_hdl)) { - rc = dbtree_iterate(root->root_hdl, DAOS_INTENT_REBUILD, false, + rc = dbtree_iterate(root->root_hdl, DAOS_INTENT_MIGRATION, false, migrate_obj_iter_cb, &arg); if (rc || tls->mpt_fini) { if (tls->mpt_status == 0)
[No CFG could be retrieved]
open the remote container as a client -DER_NONEXIST - DB - LOADED - 1.
(style) line over 80 characters
@@ -105,6 +105,15 @@ class GradeEntryFormsController < ApplicationController # For students def student_interface @grade_entry_form = GradeEntryForm.find(params[:id]) + if @grade_entry_form.is_hidden + render 'shared/http_status', formats: [:html], + locals: { + code: '404', + message: HttpStatusHelper::ERROR_CODE['message']['404'] + }, + status: 404, layout: false + return + end @student = current_user end
[GradeEntryFormsController->[new->[new],create->[new],update->[update]]]
This method is used to render a block of missing node responses for the student interface. It.
Use 2 spaces for indentation in a hash, relative to the start of the line where the left curly brace is.
@@ -1424,9 +1424,9 @@ class _SDFBoundedSourceWrapper(ptransform.PTransform): restriction_tracker=core.DoFn.RestrictionParam( _SDFBoundedSourceWrapper._SDFBoundedSourceRestrictionProvider( source, chunk_size))): - start_pos, end_pos = restriction_tracker.current_restriction() - range_tracker = self.source.get_range_tracker(start_pos, end_pos) - return self.source.read(range_tracker) + for output in restriction_tracker.get_tracking_source().read( + restriction_tracker.get_delegate_range_tracker()): + yield output return SDFBoundedSourceDoFn(self.source)
[_finalize_write->[close,finalize_write,open_writer],_SDFBoundedSourceWrapper->[_SDFBoundedSourceRestrictionProvider->[create_tracker->[_SDFBoundedSourceRestrictionTracker,get_range_tracker],split->[split],initial_restriction->[stop_position,start_position,get_range_tracker]],expand->[split_source->[],_create_sdf_bounded_source_dofn],_SDFBoundedSourceRestrictionTracker->[try_claim->[try_claim],current_restriction->[stop_position,start_position],stop_pos->[stop_position],start_pos->[start_position],try_split->[position_at_fraction,start_position,stop_pos,fraction_consumed,try_split]],_create_sdf_bounded_source_dofn->[SDFBoundedSourceDoFn->[process->[current_restriction,read,_SDFBoundedSourceRestrictionProvider,get_range_tracker]],SDFBoundedSourceDoFn,estimate_size,get_desired_chunk_size],_infer_output_coder->[default_output_coder]],_WriteKeyedBundleDoFn->[process->[close,write,open_writer]],Read->[to_runner_api_parameter->[is_bounded],from_runner_api_parameter->[Read],_infer_output_coder->[default_output_coder],expand->[split_source->[split,estimate_size,get_desired_chunk_size],get_range_tracker,read]],_WriteBundleDoFn->[finish_bundle->[close],process->[write,open_writer]],_pre_finalize->[pre_finalize],RestrictionProgress->[with_completed->[RestrictionProgress]],WriteImpl->[expand->[split_source->[],initialize_write]]]
Creates a SDF - bounded source dofn that can be used to read a sequence of.
Nit: Any reason not to simply return `restriction_tracker.get_tracking_source().read(...)`?
@@ -51,6 +51,12 @@ namespace NServiceBus.Utils.Reflection /// </summary> public static class DelegateFactory { + static readonly IDictionary<PropertyInfo, LateBoundProperty> PropertyInfoToLateBoundProperty = new ConcurrentDictionary<PropertyInfo, LateBoundProperty>(); + static readonly IDictionary<FieldInfo, LateBoundField> FieldInfoToLateBoundField = new ConcurrentDictionary<FieldInfo, LateBoundField>(); + static readonly IDictionary<PropertyInfo, LateBoundPropertySet> PropertyInfoToLateBoundPropertySet = new ConcurrentDictionary<PropertyInfo, LateBoundPropertySet>(); + static readonly IDictionary<FieldInfo, LateBoundFieldSet> FieldInfoToLateBoundFieldSet = new ConcurrentDictionary<FieldInfo, LateBoundFieldSet>(); + static readonly IDictionary<MethodInfo, LateBoundMethod> MethodInfoToLateBoundMethod = new ConcurrentDictionary<MethodInfo, LateBoundMethod>(); + /// <summary> /// Create Late Bound methods /// </summary>
[DelegateFactory->[LateBoundField->[Field,Compile,Convert,DeclaringType,Parameter,Expression],CreateParameterExpressions->[ToArray],LateBoundFieldSet->[GetILGenerator,Unbox_Any,Name,Ldarg_1,FieldType,Ldarg_0,Emit,Stfld,CreateDelegate,Castclass,DeclaringType,Ret],LateBoundProperty->[Property,Compile,Convert,DeclaringType,Parameter,Expression],LateBoundMethod->[CreateParameterExpressions,Compile,Convert,DeclaringType,Call,Parameter,Expression],LateBoundPropertySet->[GetILGenerator,Ret,Unbox_Any,Callvirt,PropertyType,Name,Ldarg_1,Ldarg_0,Emit,CreateDelegate,Castclass,DeclaringType,GetSetMethod]]]
Creates a LateBoundMethod LateBoundProperty LateBoundField and LateBound Creates a LateBoundProperty object from an instance parameter.
I may be wrong here, so I apologise in advance :smile: I don't think these need to be `ConcurrentDictionary`. #### Here is why - The dictionary items never get removed - Last one wins which is fine because the result will be the same Not making these `ConcurrentDictionary` will improve performance a little bit more :+1:
@@ -3509,6 +3509,8 @@ class Archiver: subparser.set_defaults(func=self.do_delete) subparser.add_argument('-n', '--dry-run', dest='dry_run', action='store_true', help='do not change repository') + subparser.add_argument('--list', dest='output_list', action='store_true', + help='output verbose list of items (files, dirs, ...)') subparser.add_argument('-s', '--stats', dest='stats', action='store_true', help='print statistics for the deleted archive') subparser.add_argument('--cache-only', dest='cache_only', action='store_true',
[main->[get_args,run,Archiver],with_repository->[decorator->[wrapper->[argument]]],Archiver->[do_debug_search_repo_objs->[print_error,print_finding],_export_tar->[item_to_tarinfo->[print_warning,item_content_stream],build_filter,print_warning,build_matcher,item_to_tarinfo],do_prune->[print_error],do_mount->[print_error],do_check->[print_error],do_extract->[build_filter,print_warning,build_matcher],_list_archive->[_list_inner,build_matcher],do_delete->[print_error],run->[_setup_topic_debugging,prerun_checks,get_func,_setup_implied_logging],do_debug_dump_archive->[output->[do_indent],output],do_recreate->[print_error,build_matcher],do_key_export->[print_error],do_benchmark_crud->[measurement_run,test_files],_info_archives->[format_cmdline],_process->[print_file_status,_process,print_warning],do_config->[print_error,list_config],build_parser->[define_archive_filters_group->[add_argument],define_exclusion_group->[define_exclude_and_patterns],define_borg_mount->[define_archive_filters_group,add_argument,define_exclusion_group],add_common_group,define_archive_filters_group,CommonOptions,define_borg_mount,add_argument,process_epilog,define_exclusion_group],do_key_import->[print_error],CommonOptions->[add_common_group->[add_argument->[add_argument]]],do_diff->[build_matcher,print_output,print_warning],do_debug_dump_repo_objs->[decrypt_dump],parse_args->[get_func,resolve,preprocess_args,parse_args,build_parser],do_list->[print_error],do_create->[create_inner->[print_file_status,print_error,print_warning],create_inner],with_repository],main]
Builds a parser for the given . Add options to a borg command. Add options to the borg command line. Add command line options for a specific . This function defines the command line options for Borg - Deduplicateated Backups and Command that handles the n - block block allocation and memory usage.
in this case, it is rather a list of archives, right?
@@ -1206,7 +1206,7 @@ void Planner::_buffer_line(const float &a, const float &b, const float &c, const // This leads to an enormous number of advance steps due to a huge e_acceleration. // The math is correct, but you don't want a retract move done with advance! // So this situation is filtered out here. - if (!block->steps[E_AXIS] || (!block->steps[X_AXIS] && !block->steps[Y_AXIS] && !block->steps[Z_AXIS]) || stepper.get_advance_k() == 0 || (uint32_t) block->steps[E_AXIS] == block->step_event_count) { + if (!block->steps[E_AXIS] || (!block->steps[X_AXIS] && !block->steps[Y_AXIS]) || stepper.get_advance_k() == 0 || (uint32_t) block->steps[E_AXIS] == block->step_event_count) { block->use_advance_lead = false; } else {
[No CFG could be retrieved]
Updates the current unit_vector and the nominal speed of the given block. Z - Axis.
@Sebastianv650 What's the main point here?
@@ -69,10 +69,10 @@ class RoutingRuleSchemeResource(Resource): 'allowed': ['SUN', 'MON', 'TUE', 'WED', 'THU', 'FRI', 'SAT'] }, 'hour_of_day_from': { - 'type': 'integer' + 'type': 'string' }, 'hour_of_day_to': { - 'type': 'integer' + 'type': 'string' } } }
[RoutingRuleSchemeService->[__check_if_rule_name_is_unique->[badRequestError,get,len],__validate_routing_scheme->[badRequestError,keys,get,len,format],on_update->[__check_if_rule_name_is_unique,__validate_routing_scheme],on_create->[__check_if_rule_name_is_unique,__validate_routing_scheme],on_delete->[forbiddenError,find_one]],RoutingRuleSchemeResource->[rel],getLogger]
Produces a base class for the routing_schemes endpoint. A method to check the routing scheme s rules.
This list should be declared only once and reused.
@@ -43,7 +43,7 @@ <% if user_signed_in? %> <main id="main-content"> - <header class="crayons-layout crayons-layout--limited-l crayons-layout--1-col p-4 flex items-center justify-between"> + <header class="crayons-layout crayons-layout--limited-l crayons-layout--1-col p-3 flex items-center justify-between"> <h1 class="crayons-title">Notifications</h1> <a href="<%= user_settings_path(tab: :notifications) %>" class="crayons-btn crayons-btn--ghost">Settings</a> </header>
[No CFG could be retrieved]
Displays a script that tracks clicks for welcome notifications. Renders the nagios page.
I need to componentize headers..
@@ -86,6 +86,7 @@ public class IndexIO public static final byte V8_VERSION = 0x8; public static final byte V9_VERSION = 0x9; public static final int CURRENT_VERSION_ID = V9_VERSION; + public static BitmapSerdeFactory LEGACY_FACTORY = new BitmapSerde.LegacyBitmapSerdeFactory(); public static final ByteOrder BYTE_ORDER = ByteOrder.nativeOrder();
[IndexIO->[convertSegment->[validateTwoSegments,getVersionFromDir],validateTwoSegments->[validateTwoSegments],LegacyIndexLoader->[load->[mapDir]]]]
Package private for testing purposes. Validate two segments.
This means the only way to code the presence bitmap is concise, need to add this to docs thought
@@ -214,7 +214,7 @@ public class KafkaOffsetGen { } } if (!found) { - throw new HoodieDeltaStreamerException(Config.KAFKA_AUTO_OFFSET_RESET + " config set to unknown value " + kafkaAutoResetOffsetsStr); + throw new HoodieException(Config.KAFKA_AUTO_OFFSET_RESET + " config set to unknown value " + kafkaAutoResetOffsetsStr); } if (autoResetValue.equals(KafkaResetOffsetStrategies.GROUP)) { this.kafkaParams.put(Config.KAFKA_AUTO_OFFSET_RESET.key(), Config.KAFKA_AUTO_OFFSET_RESET.defaultValue().name().toLowerCase());
[KafkaOffsetGen->[getNextOffsetRanges->[computeOffsetRanges],fetchValidOffsets->[strToOffsets],delayOffsetCalculation->[strToOffsets],commitOffsetToKafka->[strToOffsets]]]
Creates a KafkaOffsetGenerator which generates offsets for the given topic. - - - - - - - - - - - - - - - - - -.
we should leave this untouched
@@ -663,11 +663,13 @@ angular.module('ngResource', ['ng']). } if (hasBody) httpConfig.data = data; - route.setUrlParams(httpConfig, - extend({}, extractParams(data, action.params || {}), params), - action.url); - - var promise = $http(httpConfig).then(function(response) { + var promise = $q.all(extractParams(data, action.params || {})).then(function(ids) { + route.setUrlParams(httpConfig, + extend({}, ids, params), + action.url); + return $http(httpConfig); + }); + promise = promise.then(function(response) { var data = response.data; if (data) {
[No CFG could be retrieved]
Creates a resource object from the given data. Handle the response of a resource .
I don't think the first `{}` argument is needed any more. (In fact, I don't think it was needed before anyway.)
@@ -19,7 +19,7 @@ namespace System.IO.Pipelines private CancellationTokenRegistration _cancellationTokenRegistration; #if (!NETSTANDARD2_0 && !NETFRAMEWORK) - private CancellationToken CancellationToken => _cancellationTokenRegistration.Token; + private readonly CancellationToken CancellationToken => _cancellationTokenRegistration.Token; #else private CancellationToken _cancellationToken; private CancellationToken CancellationToken => _cancellationToken;
[PipeAwaitable->[Cancel->[ExtractCompletion],CancellationTokenFired->[Cancel]]]
Creates a new awaitable object that can be used to perform a Pipe operation. A helper function to register a cancellation token if it is running.
If line 22 can be readonly can't 25 also be readonly? they're the same property in different builds.
@@ -74,6 +74,8 @@ module Engine # to change order. Re-sort only them. index = @entity_index + 1 @entities[index..-1] = @entities[index..-1].sort if index < @entities.size - 1 + + @entities.pop while @entities.last.corporation? && @entities.last.share_price.liquidation? end def teleported?(entity)
[Operating->[next_entity!->[next_entity!],start_operating->[next_entity!,name]]]
Recalculates the order of the corporations in the list of entities.
would this break if entities is empty?
@@ -163,7 +163,16 @@ function LightboxWithRef( layout={true} paint={true} part="lightbox" - wrapperClassName={`${classes.defaultStyles} ${classes.wrapper}`} + contentStyle={ + // Prefer style over class to override `ContainWrapper`'s overflow + scrollable && { + overflow: 'scroll', + overscrollBehavior: 'none', + } + } + wrapperClassName={`${classes.defaultStyles} ${classes.wrapper} ${ + scrollable ? '' : classes.containScroll + }`} role="dialog" tabindex="0" onKeyDown={(event) => {
[No CFG could be retrieved]
Displays a lightbox with a single object that represents a single object that represents a single object.
Hmm. Why do we need both `contentStyle ` and `wrapperClassName ` for propagation of these two styles?
@@ -1,12 +1,15 @@ from collections import OrderedDict +from typing import Dict, cast, Type from .data_iterator import DataIterator from .basic_iterator import BasicIterator from .bucket_iterator import BucketIterator from .adaptive_iterator import AdaptiveIterator +# pylint: disable=invalid-name +iterators = OrderedDict() # type: Dict[str, Type[DataIterator]] +# pylint: enable=invalid-name -iterators = OrderedDict() # pylint: disable=invalid-name iterators["bucket"] = BucketIterator iterators["basic"] = BasicIterator iterators["adaptive"] = AdaptiveIterator
[OrderedDict]
Create an OrderedDict of the iterators for a given .
What's `cast` here? Looks unused.
@@ -76,6 +76,16 @@ public class RedisInboundChannelAdapterParserTests extends RedisAvailableTests { assertEquals(converterBean, accessor.getPropertyValue("messageConverter")); assertEquals(context.getBean("serializer"), accessor.getPropertyValue("serializer")); + Object container = accessor.getPropertyValue("container"); + DirectFieldAccessor containerAccessor = new DirectFieldAccessor(container); + Object taskExecutorObj = containerAccessor.getPropertyValue("taskExecutor"); + assertEquals(ThreadPoolTaskExecutor.class, taskExecutorObj.getClass()); + ThreadPoolTaskExecutor taskExecutor = (ThreadPoolTaskExecutor) taskExecutorObj; + assertEquals(5, taskExecutor.getCorePoolSize()); + assertEquals(10, taskExecutor.getMaxPoolSize()); + DirectFieldAccessor executorAccessor = new DirectFieldAccessor(taskExecutor); + assertEquals(25, executorAccessor.getPropertyValue("queueCapacity")); + Object bean = context.getBean("withoutSerializer.adapter"); assertNotNull(bean); assertNull(TestUtils.getPropertyValue(bean, "serializer"));
[RedisInboundChannelAdapterParserTests->[validateConfiguration->[assertNotNull,getComponentType,DirectFieldAccessor,getPropertyValue,getComponentName,assertNull,getBean,assertEquals],testInboundChannelAdapterMessaging->[assertNotNull,getPropertyValue,isOneOf,receive,awaitContainerSubscribedWithPatterns,getBean,getPayload,publish,getBytes,assertThat,getConnectionFactoryForTest],testAutoChannel->[getPropertyValue,assertSame]]]
This method validates the configuration of the bean.
Well, we don't need to assert exector's properties. There is just enough to inject that `executor` and perform `assertSame()` against `containerAccessor.getPropertyValue("taskExecutor")`
@@ -42,6 +42,9 @@ class Wheel(object): self.abis = wheel_info.group('abi').split('.') self.plats = wheel_info.group('plat').split('.') + # Adjust for any legacy custom PyPy tags found in pyversions + _add_standard_pypy_version_tags(self.pyversions) + # All the tag combinations from this file self.file_tags = { Tag(x, y, z) for x in self.pyversions
[Wheel->[get_formatted_file_tags->[sorted,str],support_index_min->[min,index],__init__->[group,InvalidWheelFilename,match,Tag],supported->[isdisjoint],compile]]
Initializes the object with the name version build_tag and file_tags from a wheel file.
Like @chrahunt suggested, in case `_add_standard_pypy_version_tags` returns `True`, `pip` should output a `PipDeprecationWarning` since we won't keep this workaround forever.
@@ -23,7 +23,7 @@ <meta name="twitter:card" content="summary_large_image"> <%= auto_discovery_link_tag(:rss, "https://dev.to/feed", title: "#{ApplicationConfig['COMMUNITY_NAME']} RSS Feed") %> <% end %> -<%= javascript_pack_tag "homePage", defer: true %> +<%= javascript_pack_tag "homePage", defer: true, "data-no-instant": true %> <% cache("main-stories-index-#{params}-#{user_signed_in?}", expires_in: 90.seconds) do %> <div class="home" id="index-container"
[No CFG could be retrieved]
Displays a list of all possible words in the system. Renders a single - item list item that is a sequence of articles.
The homepage pack bundle does not to be reevaluated. It only needs to be loaded once.
@@ -430,4 +430,10 @@ public class TestOzoneManagerPrepare extends TestOzoneManagerHA { () -> !logFilesPresentInRatisPeer(om)); } } + + private void assertShouldBeAbleToPrepare() throws Exception { + long prepareIndex = submitPrepareRequest(); + assertClusterPrepared(prepareIndex); + assertRatisLogsCleared(); + } } \ No newline at end of file
[TestOzoneManagerPrepare->[testPrepareWithTransactions->[setup],testPrepareWithRestart->[setup],testCancelPrepare->[setup],assertKeysWritten->[assertKeysWritten],assertClusterNotPrepared->[assertClusterNotPrepared],testPrepareDownedOM->[setup],testPrepareWithMultipleThreads->[setup],writeKeysAndWaitForLogs->[writeKeysAndWaitForLogs,logFilesPresentInRatisPeer],assertClusterPrepared->[assertClusterPrepared],assertRatisLogsCleared->[assertRatisLogsCleared,logFilesPresentInRatisPeer],testPrepareWithoutTransactions->[setup]]]
Assert that ratis logs are cleared.
Nit: We can use this method in testPrepareWithoutTransactions too.
@@ -62,13 +62,12 @@ function preClosureBabel() { const babel = gulpBabel({caller: {name: 'pre-closure'}}); return through.obj((file, enc, next) => { - const {relative, path} = file; - if (!filesToTransform.includes(relative)) { + if (!filesToTransform.includes(file.relative)) { return next(null, file); } - if (cache.has(path)) { - return next(null, cache.get(path)); + if (cache.has(file.path)) { + return next(null, cache.get(file.path)); } let data, err;
[No CFG could be retrieved]
A stream that transforms files before and after the compiler pass. Handle a pre - closure babel error.
Too many things with the name `path`. Good idea to not spread here since it's getting confusing which `path` is what and where. Perhaps in the future an ESLint rule for `build-system` prohibiting the spread of something creating a `path` variable might make sense.
@@ -3,6 +3,8 @@ class OrganizationsController < ApplicationController rescue_from Errno::ENAMETOOLONG, with: :log_image_data_to_datadog def create + rate_limit! + @tab = "organization" @user = current_user @tab_list = @user.settings_tab_list
[OrganizationsController->[organization_params->[strip_tags,transform_values,name],create->[redirect_to,id,new,render,authorize,create!,save,valid_filename?,settings_tab_list],set_organization->[authorize,find_by],update->[redirect_to,render,merge,update,current,valid_filename?,settings_tab_list],generate_new_secret->[redirect_to,save,secret,generated_random_secret],valid_filename?->[new,dig,authorize,long_filename?,add,except],after_action,rescue_from]]
create a new lease file.
Any reason to not put this in a `before_action`? Seems like we wouldn't even want to continue with this request if the rate limiter decides that we've exceeded the limit?
@@ -221,6 +221,11 @@ ReturnFalse: { if (ClassType == ClassType.Object) { + if (ExtensionDataPropertyTypeSpec != null) + { + return false; + } + foreach (PropertyGenerationSpec property in PropertyGenSpecList) { if (property.TypeGenerationSpec.Type.IsObjectType() ||
[TypeGenerationSpec->[FastPathIsSupported->[Object,IDictionaryOfTKeyTValue,IDictionary,AllowNamedFloatingPointLiterals,ConverterInstantiationLogic,Dictionary,IReadOnlyDictionary,IsStringType,WriteAsString,IsObjectType,NumberHandling,ImmutableDictionary,NotApplicable],Initialize->[GetTypeInfoPropertyName,GetCompilableName,IsValueType],TryFilterSerializableProps->[Always,IsPublic,ClrName,HasJsonInclude,Add,Count,DefaultIgnoreCondition,RuntimePropertyName,Assert,IncludeFields,WhenWritingNull,IsProperty,CanBeNull,TryAdd],GenerationModeIsSpecified->[Default],GetImmutableEnumerableConstructingTypeName,ImmutableEnumerable,CreateRangeMethodName,Metadata,GetImmutableDictionaryConstructingTypeName,Assert,FastPathIsSupported,Serialization,GenerationModeIsSpecified,ImmutableDictionary]]
Fast path is supported.
Why can't the FastPath support this?
@@ -243,7 +243,6 @@ public class HttpTemplateDownloader extends ManagedContextRunnable implements Te offset = writeBlock(bytes, out, block, offset); if (!verifyFormat.isVerifiedFormat() && (offset >= 1048576 || offset >= remoteSize)) { //let's check format after we get 1MB or full file verifyFormat.invoke(); - if (verifyFormat.isInvalid()) return true; } } else { done = true;
[HttpTemplateDownloader->[stopDownload->[getStatus],getDownloadLocalPath->[getToFile],runInContext->[download],main->[getDownloadTime,download,HttpTemplateDownloader]]]
Copy bytes from in to out.
@Spaceman1984 cc @nvazquez why are we not validating format now?
@@ -260,6 +260,11 @@ final class InternalExternalizers { addInternalExternalizer(new SegmentPublisherResult.Externalizer(), exts); addInternalExternalizer(new PublisherReducers.PublisherReducersExternalizer(), exts); addInternalExternalizer(new CacheStreamIntermediateReducer.ReducerExternalizer(), exts); + addInternalExternalizer(new ClassExternalizer(gcr.getGlobalConfiguration().classLoader()), exts); + addInternalExternalizer(new ClusterCacheStatsImpl.DistributedCacheStatsCallableExternalizer(), exts); + addInternalExternalizer(ThrowableExternalizer.INSTANCE, exts); + addInternalExternalizer(new ImmutableListCopy.Externalizer(), exts); + addInternalExternalizer(EnumExternalizer.INSTANCE, exts); return exts; }
[InternalExternalizers->[addInternalExternalizer->[getTypeClasses,put],load->[IntSetExternalizer,CollectionExternalizer,ImmutableSetWrapperExternalizer,NoValueReadOnlyViewExternalizer,MapOpsExternalizer,UuidExternalizer,EquivalenceExternalizer,CacheFiltersExternalizer,AdminFlagExternalizer,ImmutableMapWrapperExternalizer,TerminalOperationExternalizer,XSiteStateExternalizer,ReadWriteSnapshotViewExternalizer,DoubleSummaryStatisticsExternalizer,LongSummaryStatisticsExternalizer,SetValueIfEqualsReturnBooleanExternalizer,LambdaWithMetasExternalizer,getComponent,EntryVersionParamExternalizer,addInternalExternalizer,IntSummaryStatisticsExternalizer,ReducerExternalizer,ReadOnlySnapshotViewExternalizer,CacheRpcCommandExternalizer,EnumSetExternalizer,XidExternalizer,TriangleAckExternalizer,IteratorResponsesExternalizer,ReplicableCommandExternalizer,MapExternalizer,ClassToExternalizerMap,PublisherReducersExternalizer,IntermediateOperationExternalizer,EndIteratorExternalizer,OptionalExternalizer,Externalizer,MaxIdleExternalizer,ConstantLambdaExternalizer,StreamMarshallingExternalizer,LifespanExternalizer]]]
Loads a cache externalizer map. Adds all internal externalizers to the internal set. Adds all internal externalizers to the cache. Add external classes to the internal list. Adds an externalizer to the externalizer map.
Looks like we can remove `ImmutableListCopyExternalizer` as well now?
@@ -82,11 +82,11 @@ echo "</td><td width=100 onclick=\"location.href='" . generate_port_url($port) . if ($port_details) { $port['graph_type'] = 'port_bits'; - echo generate_port_link($port, "<img src='graph.php?type=port_bits&amp;id=" . $port['port_id'] . '&amp;from=' . Config::get('time.day') . '&amp;to=' . Config::get('time.now') . '&amp;width=100&amp;height=20&amp;legend=no&amp;bg=' . str_replace('#', '', $row_colour) . "'>"); + echo generate_port_link($port, "<img src='graph.php?type=port_bits&amp;id=" . $port['port_id'] . '&amp;from=' . Config::get('time.day') . '&amp;to=' . Config::get('time.now') . '&amp;width=100&amp;height=20&amp;legend=no&amp;bg=' . str_replace('#', '', $row_colour) . "00'>"); $port['graph_type'] = 'port_upkts'; - echo generate_port_link($port, "<img src='graph.php?type=port_upkts&amp;id=" . $port['port_id'] . '&amp;from=' . Config::get('time.day') . '&amp;to=' . Config::get('time.now') . '&amp;width=100&amp;height=20&amp;legend=no&amp;bg=' . str_replace('#', '', $row_colour) . "'>"); + echo generate_port_link($port, "<img src='graph.php?type=port_upkts&amp;id=" . $port['port_id'] . '&amp;from=' . Config::get('time.day') . '&amp;to=' . Config::get('time.now') . '&amp;width=100&amp;height=20&amp;legend=no&amp;bg=' . str_replace('#', '', $row_colour) . "00'>"); $port['graph_type'] = 'port_errors'; - echo generate_port_link($port, "<img src='graph.php?type=port_errors&amp;id=" . $port['port_id'] . '&amp;from=' . Config::get('time.day') . '&amp;to=' . Config::get('time.now') . '&amp;width=100&amp;height=20&amp;legend=no&amp;bg=' . str_replace('#', '', $row_colour) . "'>"); + echo generate_port_link($port, "<img src='graph.php?type=port_errors&amp;id=" . $port['port_id'] . '&amp;from=' . Config::get('time.day') . '&amp;to=' . Config::get('time.now') . '&amp;width=100&amp;height=20&amp;legend=no&amp;bg=' . str_replace('#', '', $row_colour) . "00'>"); } echo "</td><td width=120 onclick=\"location.href='" . generate_port_url($port) . "'\" >";
[toArray,hasGlobalRead]
Print a list of ports. Displays the link to the port in the network.
how does this affect light theme?
@@ -894,7 +894,7 @@ class DygraphGeneratorLoader(DataLoaderBase): array = core.LoDTensorArray() for item in sample: if not isinstance(item, core.LoDTensor): - self._check_input_array(item) + item = self._check_input_array(item) tmp = core.LoDTensor() tmp.set(item, core.CPUPlace()) item = tmp
[PyReader->[__iter__->[__iter__],__next__->[__next__],reset->[reset],decorate_batch_generator->[set_batch_generator],__init__->[from_generator],start->[start],decorate_sample_list_generator->[set_sample_list_generator],decorate_sample_generator->[set_sample_generator]],DygraphGeneratorLoader->[__iter__->[_init_iterable,_start],__next__->[_reset],_init_iterable->[_wait_process_ends,_wait_thread_ends],_start->[__thread_main__->[],_clear_and_remove_data_queue],_reader_thread_loop_for_singleprocess->[_check_input_array],set_batch_generator->[_convert_places],_reset->[_wait_process_ends,_wait_thread_ends],_reader_thread_loop_for_multiprocess->[_exit_thread_unexpectedly,_exit_thread_expectedly]],DatasetLoader->[__iter__->[_start],__init__->[_convert_places]],DataLoader->[__call__->[__iter__],__init__->[_convert_places]],GeneratorLoader->[__iter__->[_init_iterable,_start],__next__->[_reset],set_sample_list_generator->[set_batch_generator],_init_iterable->[_wait_thread_ends],_start->[__thread_main__->[_check_input_array],start],reset->[_reset],set_batch_generator->[_convert_places],__init__->[keep_data_loader_order],start->[_start],set_sample_generator->[set_sample_list_generator,set_batch_generator],_reset->[reset]]]
This loop is called from the reader thread.
Is there any efficiency difference between setting list or ndarray to a LoDTensor ?
@@ -14,9 +14,14 @@ limitations under the License. """ +import sys +from pathlib import Path +sys.path.append(str(Path(__file__).resolve().parents[3] / 'common/python/models')) + from collections import namedtuple import cv2 import numpy as np +from utils import nms ModelAttributes = namedtuple('ModelAttributes', ['required_outputs', 'postprocessor'])
[mask_rcnn_postprocess->[segm_postprocess],segm_postprocess->[expand_box],yolact_segm_postprocess->[crop_mask->[sanitize_coordinates],sanitize_coordinates,crop_mask]]
Expands a box to a 2D array with the specified number of non - zero values.
This shouldn't be needed if the main script of the demo already modifies `sys.path` (which it does).
@@ -129,7 +129,7 @@ public class SparkExecuteClusteringCommitActionExecutor<T extends HoodieRecordPa /** * Validate actions taken by clustering. In the first implementation, we validate at least one new file is written. * But we can extend this to add more validation. E.g. number of records read = number of records written etc. - * + * * We can also make these validations in BaseCommitActionExecutor to reuse pre-commit hooks for multiple actions. */ private void validateWriteResult(HoodieWriteMetadata<JavaRDD<WriteStatus>> writeMetadata) {
[SparkExecuteClusteringCommitActionExecutor->[buildWriteMetadata->[getPartitionToReplacedFileIds]]]
Validate the write result.
I have no idea how these are getting included. I reverted this locally and did git add of this file. I pushed it and repeated it twice, but it gets auto formatted somehow :(
@@ -78,7 +78,7 @@ from apache_beam.utils import processes # Update this version to the next version whenever there is a change that will # require changes to legacy Dataflow worker execution environment. # This should be in the beam-[version]-[date] format, date is optional. -BEAM_CONTAINER_VERSION = 'beam-2.2.0-20170807' +BEAM_CONTAINER_VERSION = 'beam-2.2.0-20170928' # Update this version to the next version whenever there is a change that # requires changes to SDK harness container or SDK harness launcher. # This should be in the beam-[version]-[date] format, date is optional.
[get_sdk_package_name->[get_sdk_name_and_version],stage_job_resources->[_stage_extra_packages],_build_setup_package->[_get_python_executable],get_sdk_name_and_version->[_get_required_container_version],_download_pypi_sdk_package->[get_sdk_package_name,_get_python_executable],_stage_beam_sdk_tarball->[_dependency_file_download,_dependency_file_copy],_populate_requirements_cache->[_get_python_executable],_stage_extra_packages->[_dependency_file_copy]]
Imports a single - version object from the main system. Copies a local file to a GCS file or vice versa.
<!--new_thread; commit:87ea0681e36c0640c02c8d137f83ad73b95e1bc3; resolved:0--> The point of using a dated version string is that incompatibilities should not break users who check out the beam repo. We should rebuild a new container at the current date instead of using the catch-all dev container.
@@ -195,6 +195,8 @@ public class PipelineMessageNotificationTestCase extends AbstractReactiveProcess event = Event.builder(context).message(InternalMessage.builder().payload("request").build()).exchangePattern(ONE_WAY) .flow(pipeline).build(); + thrown.expect(instanceOf(MessagingException.class)); + thrown.expectCause(instanceOf(IllegalStateException.class)); try { processFlow(pipeline, event); } finally {
[PipelineMessageNotificationTestCase->[TestPipeline->[configureMessageProcessors->[configureMessageProcessors],configurePostProcessors->[configurePostProcessors]]]]
This method is called when a message is received from a pipeline that is one way.
what's the reason of this change? could a message assertion be added too?
@@ -75,9 +75,6 @@ public class RocksDBDAO { * Create RocksDB if not initialized. */ private RocksDB getRocksDB() { - if (null == rocksDB) { - init(); - } return rocksDB; }
[RocksDBDAO->[dropColumnFamily->[dropColumnFamily,get],close->[close],prefixDelete->[delete,get],addColumnFamily->[put,getColumnFamilyDescriptor],get->[get],put->[put],delete->[delete],prefixSearch->[get]]]
Get the RocksDB object.
What if getRocksDB is called repetitively ? It should not init() every time, do we plan to add a different check in init to allow for better code coverage ?
@@ -497,7 +497,7 @@ class LocalCache(CacheStatsMixin): self.cache_config.create() ChunkIndex().write(os.path.join(self.path, 'chunks')) os.makedirs(os.path.join(self.path, 'chunks.archive.d')) - with SaveFile(os.path.join(self.path, 'files'), binary=True): + with SaveFile(os.path.join(self.path, files_cache_name()), binary=True): pass # empty file def _do_open(self):
[assert_secure->[SecurityManager,assert_secure],Cache->[destroy->[exists],break_lock->[cache_dir],__new__->[CacheConfig,adhoc,exists,local]],CacheConfig->[load->[recanonicalize_relative_location,open],create->[exists],__init__->[cache_dir],exists->[exists],_check_upgrade->[close]],AdHocCache->[commit->[save],chunk_decref->[begin_txn],seen_chunk->[begin_txn],__init__->[SecurityManager,assert_secure],chunk_incref->[begin_txn],add_chunk->[chunk_incref,begin_txn,seen_chunk]],SecurityManager->[assert_location_matches->[save],assert_key_type->[known,key_matches],_assert_secure->[assert_location_matches,assert_key_type,assert_no_manifest_replay,known,save],assert_access_unknown->[known,save],key_matches->[known]],CacheStatsMixin->[chunks_stored_size->[stats],format_tuple->[stats]],LocalCache->[sync->[read_archive_index->[write_archive_index,mkpath,cleanup_cached_archive],cleanup_cached_archive->[mkpath],write_archive_index->[mkpath],create_master_idx->[fetch_missing_csize,cleanup_outdated,read_archive_index,get_archive_ids_to_names,repo_archives,cached_archives,fetch_and_build_idx],fetch_and_build_idx->[fetch_missing_csize],create_master_idx,legacy_cleanup,begin_txn],_do_open->[load],commit->[save],create->[create,open],__init__->[cache_dir,create,close,CacheConfig,exists,SecurityManager,assert_secure,assert_access_unknown,open],rollback->[exists,_do_open],chunk_decref->[begin_txn],chunk_incref->[begin_txn],add_chunk->[begin_txn],close->[close],__exit__->[close],open->[open]]]
Create a new empty cache at self. path.
did you systematically search for all occurances of `'files'` or `"files"` and replace them all?
@@ -83,7 +83,7 @@ def test_lcmv(): stc_normal = lcmv(evoked, forward_surf_ori, noise_cov, data_cov, reg=0.01, pick_ori="normal") - assert_true((np.abs(stc_normal.data) <= stc.data + 0.8).all()) + assert_true((np.abs(stc_normal.data) <= stc.data + 0.4).all()) # Test picking source orientation maximizing output source power stc_max_power = lcmv(evoked, forward, noise_cov, data_cov, reg=0.01,
[test_lcmv_raw->[lcmv_raw,assert_array_almost_equal,read_selection,read_cov,regularize,assert_true,len,time_as_index,pick_types,compute_raw_data_covariance,intersect1d,array],test_lcmv->[,lcmv,max,read_cov,sum,assert_raises,argmax,resample,assert_array_almost_equal,drop_bad_epochs,dict,regularize,compute_covariance,zeros_like,stcs,assert_array_equal,read_selection,len,lcmv_epochs,next,average,assert_true,Epochs,pick_types],Raw,read_events,read_label,read_cov,read_forward_solution,data_path,join]
Test LCMV with evoked data and single trials Test if a single trial is detected when picking a specific block of data. Checks if the sequence of source vertices in the list of sources has a .
Why can it be bigger?
@@ -106,6 +106,8 @@ public class MustFightBattle extends DependentBattle private static final long serialVersionUID = 5879502298361231540L; + private static final long MAX_INFINITE_ROUNDS = 10000; + private final Collection<Unit> attackingWaitingToDie = new ArrayList<>(); private final Collection<Unit> defendingWaitingToDie = new ArrayList<>();
[MustFightBattle->[showCasualties->[isEmpty],remove->[isEmpty],execute->[execute],getBattleExecutables->[updateDefendingAaUnits,updateOffensiveAaUnits],getAttackerRetreatTerritories->[isEmpty,remove],isEmpty->[isEmpty],determineStepStrings->[updateDefendingAaUnits,updateOffensiveAaUnits],checkDefendingPlanesCanLand->[isEmpty,remove],newFiringUnitGroups->[isEmpty],fireNavalBombardment->[isEmpty],unitsLostInPrecedingBattle->[isEmpty],attackerWins->[endBattle],endBattle->[isEmpty],markAttackingTransports->[isEmpty],removeFromDependents->[unitsLostInPrecedingBattle],removeCasualties->[isEmpty],addRoundResetStep->[execute->[determineStepStrings,updateDefendingAaUnits,pushFightLoopOnStack,isEmpty,updateOffensiveAaUnits]],damagedChangeInto->[isEmpty],retreatQuery->[retreatQuery],fire->[isEmpty],addPlayerCombatHistoryText->[isEmpty],querySubmergeTerritory->[retreatQuery],removeFromNonCombatLandings->[isEmpty,remove],fight->[addDependentUnits,endBattle,isEmpty],defenderWins->[endBattle,isEmpty],removeNonCombatants->[isEmpty,removeNonCombatants],findTargetGroupsAndFire->[fire],nobodyWins->[endBattle],getUnits->[getUnits]]]
A base class for all the battles that need to be fighted. Creates a MustFightBattle object.
What do you think of naming this `MAX_ROUNDS` instead? "max infinite" does not quite grok well for me.
@@ -53,7 +53,7 @@ class FreqtradeBot(object): self.rpc: RPCManager = RPCManager(self) - exchange_name = self.config.get('exchange', {}).get('name', 'bittrex').title() + exchange_name = self.config.get('exchange', {}).get('name').title() self.exchange = ExchangeResolver(exchange_name, self.config).exchange self.wallets = Wallets(self.config, self.exchange)
[FreqtradeBot->[execute_buy->[get_target_bid,_get_min_pair_stake_amount],cleanup->[cleanup],process_maybe_execute_buy->[create_trade],update_trade_state->[get_real_amount],handle_timedout_limit_buy->[handle_buy_order_full_cancel],handle_trade->[get_sell_rate],create_trade->[_get_trade_stake_amount],notify_sell->[get_sell_rate]]]
Initialize the object with all variables and objects.
Using the default here was pointless anyway, since it was in the "required" fields and not having it wouldn't have passed validation...
@@ -109,7 +109,15 @@ export class Xhr { /** @const {!Window} */ this.win = win; + /** @private */ const ampdocService = Services.ampdocServiceFor(win); + + // The isSingleDoc check is required because if in shadow mode, this will + // throw a console error because the shellShadowDoc_ is not set when + // fetching the amp doc. So either the test-bind-impl or test pre setup in + // shadow mode tests needs to be fixed or there is a bug in ampdoc impl + // getAmpDoc. + // TODO(alabiaga): This should be investigated and fixed /** @private {?./ampdoc-impl.AmpDoc} */ this.ampdocSingle_ = ampdocService.isSingleDoc() ? ampdocService.getAmpDoc() : null;
[No CFG could be retrieved]
A service that polyfills Fetch API for use within AMP. If the request is a native URL and the browser supports the then it will intercept.
This isn't a property, so can't be `@private`.
@@ -230,4 +230,5 @@ public interface StorageContainerLocationProtocol extends Closeable { */ boolean getReplicationManagerStatus() throws IOException; + List<String> getScmRatisStatus() throws IOException; }
[No CFG could be retrieved]
Returns the replication manager status.
Does this API belongs to `StorageContainerLocationProtocol`?
@@ -38,7 +38,7 @@ type ComplianceIngestServer struct { mgrClient manager.NodeManagerServiceClient automateURL string notifierClient notifier.Notifier - updateManager *projectupdater.Manager + updateManager *project_update_lib.DomainProjectUpdateManager } var MinimumSupportedInspecVersion = semver.MustParse("2.0.0")
[handleNotifications->[Warnf,Compliance,Send,Errorf,WithFields],ProjectUpdateStatus->[PercentageComplete,EstimatedTimeCompelete,Errorf,TimestampProto,State],ProcessComplianceReport->[handleNotifications,GetRelease,Error,LTE,TimestampProto,sendNodeInfoToManager,GetName,Errorf,Debugf,Parse,Make,Run,FromString,Debug],HandleEvent->[Start,Errorf,Cancel,Debugf],sendNodeInfoToManager->[Wrap,ProcessNode,Errorf,Debugf],NewCompliancePipeline,NewManager,GetStringValue,Errorf,MustParse]
NewComplianceIngestServer creates a new ComplianceIngestServer object. HandleEvent creates a new instance of ComplianceIngestServer.
Removed the specific compliance project update manager to use the common domain project update manager
@@ -38,6 +38,8 @@ import java.util.stream.Collectors; public class VirtManagerSalt implements VirtManager { private final SaltApi saltApi; + private MinionPillarFileManager minionVirtualizationPillarFileManager = + new MinionPillarFileManager(new MinionVirtualizationPillarGenerator()); /** * Service providing utility functions to handle virtual machines.
[VirtManagerSalt->[getGuestDefinition->[of,parse,empty,callSync,put,map],getCapabilities->[empty,callSync],getNetworks->[toMap,of,getKey,orElse,empty,callSync,collect,getAsJsonObject],getPoolCapabilities->[empty,callSync],getVolumes->[of,orElse,empty,callSync,asList],getPools->[toMap,of,getKey,orElse,empty,callSync,collect,getAsJsonObject],getPoolDefinition->[of,setAutostart,getAsInt,parse,empty,callSync,put,getPools,map],updateLibvirtEngine->[singletonList,of,apply,hasVirtualizationEntitlement,callSync,put,getMinionId]]]
Creates a new virtual machine definition. Gets the Guest capabilities for a given minion.
Shouldn't this be called "SaltVirtManager" ?
@@ -88,6 +88,9 @@ public abstract class ExposedConfiguration { @JsonProperty("gc_warning_threshold") public abstract String gcWarningThreshold(); + @JsonProperty("forwarder_enabled") + public abstract boolean forwarderEnabled(); + public static ExposedConfiguration create(Configuration configuration) { return create( configuration.getInputbufferProcessors(),
[ExposedConfiguration->[create->[getInputbufferProcessors,getBinDir,create,getStreamProcessingMaxFaults,toString,isAllowHighlighting,getInputBufferRingSize,isAllowLeadingWildcardSearches,getNodeIdFile,getOutputModuleTimeout,getStreamProcessingTimeout,getDataDir,getName,getPluginDir,getProcessBufferProcessors,getRingSize,getStaleMasterTimeout,getOutputBufferProcessors,AutoValue_ExposedConfiguration]]]
Creates an exposed configuration object from a given configuration object.
Do we actually need to expose that flag in the configuration API? I think just having `isForwarderEnabled()` should be fine
@@ -413,7 +413,7 @@ public class AnnotationsBasedDescriberTestCase extends AbstractAnnotationsBasedD } private void assertTestModuleOperations(ExtensionDeclaration extensionDeclaration) throws Exception { - assertThat(extensionDeclaration.getOperations(), hasSize(22)); + assertThat(extensionDeclaration.getOperations(), hasSize(23)); WithOperationsDeclaration withOperationsDeclaration = extensionDeclaration.getConfigurations().get(0); assertThat(withOperationsDeclaration.getOperations().size(), is(8));
[AnnotationsBasedDescriberTestCase->[listOfString->[build],minMuleVersionIsDescribedCorrectly->[getDeclaration,describerFor,MuleVersion,setDescriber,getMinMuleVersion,is,assertThat,describeExtension],messageOperationWithoutGenerics->[getDeclaration,assertThat,describe,DefaultDescribingContext,getType,is,instanceOf,getClassLoader,getOperation],assertModelProperties->[get,notNullValue,getType,is,assertThat,isAssignableFrom],categoryDefaultValueIsDescribedCorrectly->[getDeclaration,describerFor,setDescriber,getCategory,is,assertThat,describeExtension],interceptingOperationWithoutAttributes->[getDeclaration,describerFor,assertOutputType,build,getOutputAttributes,getOutput,setDescriber,notNullValue,toMetadataType,getConfiguration,is,assertThat,getOperation,describeExtension],heisenbergWithOperationsConfig->[describe,getClassLoader,DefaultDescribingContext],assertTestModuleOperations->[getExceptionEnricherFactory,arrayOf,size,getType,hasSize,nullValue,getOperations,getOperation,toMetadataType,empty,isEmpty,load,isPresent,createEnricher,objectTypeBuilder,equalTo,notNullValue,assertOperation,is,assertParameter,build,get,getParameters,instanceOf,assertThat],interceptingOperationWithAttributes->[getDeclaration,describerFor,assertOutputType,getOutputAttributes,getOutput,setDescriber,notNullValue,toMetadataType,getConfiguration,is,assertThat,getOperation,describeExtension],setUp->[describerFor,setDescriber],heisenbergPointerPlusExternalConfig->[getDeclaration,describerFor,size,getName,equalTo,get,setDescriber,notNullValue,getParameters,toMetadataType,hasSize,is,assertParameter,assertThat,assertExtensionProperties],assertExtensionProperties->[getName,getDescription,notNullValue,is,assertThat,getVersion],heisenbergWithParameterGroupAsOptional->[describe,getClassLoader,DefaultDescribingContext],heisenbergWithRecursiveParameterGroup->[describe,getClassLoader,DefaultDescribingContext],heisenbergWithOperationPointingToExtension->[describe,getClassLoader,DefaultDescribingContext],minMuleVersionDefaultValueIsDescribedCorrectly->[getDeclaration,describerFor,MuleVersion,setDescriber,getMinMuleVersion,is,assertThat,describeExtension],heisenbergWithOperationPointingToExtensionAndDefaultConfig->[describe,getClassLoader,DefaultDescribingContext],heisenbergPointer->[getDeclaration,assertTestModuleMessageSource,describerFor,setDescriber,assertTestModuleConfiguration,assertTestModuleOperations,assertModelProperties,assertTestModuleConnectionProviders,assertExtensionProperties,describeExtension],assertOperation->[getDescription,equalTo,notNullValue,is,assertThat,getOperation],assertTestModuleConfiguration->[listOfString,arrayOf,getName,build,objectTypeBuilder,get,equalTo,getParameters,toMetadataType,hasSize,getConfigurations,assertParameter,assertThat,id],assertParameter->[getName,getDescription,equalTo,notNullValue,getType,findParameter,is,getDefaultValue,isRequired,assertThat,getExpressionSupport],assertTestModuleConnectionProviders->[getName,get,equalTo,getParameters,toMetadataType,getType,hasSize,is,assertParameter,assertThat,getConnectionProviders],describeTestModule->[getDeclaration,assertTestModuleMessageSource,assertTestModuleConfiguration,assertTestModuleOperations,assertModelProperties,assertTestModuleConnectionProviders,assertExtensionProperties,describeExtension],assertTestModuleMessageSource->[getName,get,equalTo,getParameters,toMetadataType,getType,hasSize,is,getMessageSources,assertParameter,assertThat],heisenbergWithMoreThanOneConfigInOperation->[describe,getClassLoader,DefaultDescribingContext],assertOutputType->[equalTo,hasDynamicType,getType,is,assertThat],flyweight->[getDeclaration,describerFor,assertThat,getSimpleName,setDescriber,is,getMessageSources,getConfigurations,getOperations,findDeclarationByName,sameInstance,describeExtension],findDeclarationByName->[NoSuchElementException,orElseThrow],categoryIsDescribedCorrectly->[getDeclaration,describerFor,setDescriber,getCategory,is,assertThat,describeExtension],getProductVersion]]
Checks that the module operations are present in the extension. This method checks if the operation is in the list of operations. This method checks if the parameters of the operation are not null. region ExtensionManager Implementation Get the most specific operation in the extension.
each of those 22 other operations have assertions. This one should too
@@ -2815,4 +2815,16 @@ def test_file_like(kind, preload, tmpdir): assert file_fid.closed +def test_epochs_get_data_item(): + """Test epochs.get_data(item=...).""" + raw, events, _ = _get_data() + epochs = Epochs(raw, events[:10], event_id, tmin, tmax) + with pytest.raises(ValueError, match='item must be None'): + epochs.get_data(item=0) + epochs.drop_bad() + one_data = epochs.get_data(item=0) + one_epo = epochs[0] + assert_array_equal(one_data, one_epo.get_data()) + + run_tests_if_main()
[test_epochs_bad_baseline->[_get_data],test_drop_channels_mixin->[_get_data],test_average_methods->[fun],test_event_ordering->[_get_data],test_epochs_io_preload->[_get_data],test_channel_types_mixin->[_get_data],test_resample->[_get_data],test_shift_time->[_get_data],test_detrend->[_get_data],test_illegal_event_id->[_get_data],test_epochs_proj->[_get_data],test_drop_epochs_mult->[_get_data],test_epoch_eq->[_get_data],test_decim->[_get_data],test_epoch_combine_ids->[_get_data],test_readonly_times->[_get_data],test_base_epochs->[_get_data],test_reject->[_get_data],test_evoked_arithmetic->[_get_data],test_epochs_io_proj->[_get_data],test_to_data_frame->[_get_data],test_delayed_epochs->[_get_data],test_access_by_name->[_get_data],test_concatenate_epochs_large->[_get_data],test_shift_time_raises_when_not_loaded->[_get_data],test_save_complex_data->[_get_data],test_io_epochs_basic->[_get_data],test_add_channels_epochs->[make_epochs,_get_data],test_preload_epochs->[_get_data],test_epochs_hash->[_get_data],test_subtract_evoked->[_get_data],test_epochs_drop_selection->[_get_selection],test_reject_epochs->[_get_data],test_epochs_proj_mixin->[_get_data],test_evoked_io_from_epochs->[_get_data],test_pick_channels_mixin->[_get_data],test_hierarchical->[_get_data],test_comparision_with_c->[_get_data],test_concatenate_epochs->[_get_data],test_equalize_channels->[_get_data],test_savgol_filter->[_get_data],test_add_channels->[_get_data],test_filter->[_get_data],test_epochs_copy->[_get_data],test_drop_epochs->[_get_data],test_indexing_slicing->[_get_data],test_epoch_multi_ids->[_get_data],test_read_epochs_bad_events->[_get_data],test_own_data->[_get_data],test_crop->[_get_data],test_default_values->[_get_data],test_bootstrap->[_get_data],test_events_type->[_get_data],test_iter_evoked->[_get_data],test_evoked_standard_error->[_get_data],test_contains->[_get_data]]
This is a test method that should be called only once.
can you parametrize the test so you test both preload = True or False?
@@ -8,7 +8,8 @@ module GobiertoPlans before_action :load_plan_types, only: [:index] def index - @plan_type = @plan_types.last + # HACK: select last plan which at least has one published plan to avoid https://github.com/PopulateTools/issues/issues/304 + @plan_type = @plan_types.select{ |plan_type| plan_types.plans.published.any? }.last load_plans load_years @year = @years.first
[PlanTypesController->[load_plans->[published],show->[call,redirect_to,nil?,new,respond_to,render,count,configuration_data,html,json,gobierto_plans_plan_path,first,levels,plan_updated_at],load_years->[reverse!],load_year->[to_i],load_plan_types->[sort_by_updated_at],find_plan_type->[find_by!],find_plan->[find_by!],index->[redirect_to,slug,gobierto_plans_plan_path,first,last],level_keys->[select],before_action,include]]
404 - user has no lease for a particular plan type.
Space missing to the left of {.<br>Unused block argument - plan_type. You can omit the argument if you don't care about it.
@@ -31,6 +31,11 @@ function stubAsyncErrorThrows() { rethrowAsyncSandbox = sinon.createSandbox(); rethrowAsyncSandbox.stub(coreError, 'rethrowAsync').callsFake((...args) => { const error = coreError.createErrorVargs.apply(null, args); + const index = indexOfExpectedMessage(error.message); + if (index != -1) { + expectedAsyncErrors.splice(index, 1); + return; + } self.__AMP_REPORT_ERROR(error); throw error; });
[No CFG could be retrieved]
Creates a stub and a restore function that will throw an error when a test fails.
Clever! This will prevent errors from leaking from tests, and could really improve reliability. /cc @rileyajones as an FYI.
@@ -184,7 +184,7 @@ func buildCmd(c *cli.Context) error { if err != nil { return errors.Wrapf(err, "error building system context") } - + systemContext.AuthFilePath = getAuthFile(c.String("authfile")) commonOpts, err := parse.CommonBuildOptions(c) if err != nil { return err
[RemoveAll,Dir,IsRootless,NamespaceOptions,StringSlice,IDMappingOptions,BoolT,Close,LookupEnv,HasPrefix,IsSet,UseLayers,Args,SystemContextFromOptions,GetRuntime,ValidateFlags,Errorf,TempDirForURL,Bool,SplitN,Wrapf,Join,Tail,Shutdown,Rel,Build,GlobalString,CommonBuildOptions,SetOutput,String,AddOrReplace,Abs,OpenFile]
if parses command line options and builds a single ID. The command line interface for the signature - policy command.
nit of the tiniest nits. If possible, it would be nice to have the blank line above this still. But only adjust if you need other refinements for this PR.
@@ -1417,6 +1417,8 @@ func bindSpec(spec PackageSpec, languages map[string]Language, loader Loader) (* } else { version = &v } + } else { + diags = diags.Append(errorf("#/version", "No version provided")) } // Parse the module format, if any.
[marshalResource->[marshalObjectData,marshalObject],bindProperties->[newOptionalType,bindType],bindResourceType->[bindResourceTypeDetails],MarshalYAML->[MarshalSpec],bindEnumTypeDetails->[bindPrimitiveType],marshalType->[marshalType],bindObjectType->[bindObjectTypeDetails],marshalFunction->[marshalObject],bindType->[newMapType,bindPrimitiveType,bindType,bindTypeSpecRef,newArrayType,newInputType,newUnionType],MarshalSpec->[String,IsInputShape],MarshalJSON->[MarshalSpec],bindTypeSpecRef->[GetType,bindPrimitiveType,GetResourceType,parseTypeSpecRef],bindEnumType->[bindEnumTypeDetails],marshalObject->[marshalObjectData],bindObjectTypeDetails->[bindProperties],String->[String],newUnionType->[String],bindProperties,bindEnumTypeDetails,bindObjectType,ImportLanguages,bindObjectTypeDetails,String]
bindSpec is the main entry point for the binding of a package type. It returns the if returns a context for the next generation of the plugin.
This will not work without downstream code changes. We allow schemas to elide the version argument to handle situations where the version is not known at the time the schema is generated, which is the case in ~all of our providers. All of our providers pull their version from a git tag, which is not available at the time the schema is generated; it's complicated.
@@ -92,6 +92,17 @@ def _prepare_gain(forward, info, noise_cov, pca, depth, loose, weights, raise ValueError('Depth parameter must be positive (got %s).' % depth) + if len(forward['src']) == 1: + loose = 1.0 + logger.info('Volume source space: setting loose = 1.') + if isinstance(weights, SourceEstimate): + raise ValueError('Weights must not be a cortically constrained ' + 'source estimate.') + else: + if isinstance(weights, VolSourceEstimate): + raise ValueError('Weights must not be a volume-based ' + 'source estimate.') + gain, gain_info, whitener, source_weighting, mask = \ _prepare_gain_column(forward, info, noise_cov, pca, depth, loose, weights, weights_min)
[tf_mixed_norm->[_reapply_source_weighting,_prepare_gain,_make_sparse_stc,_compute_residual,_window_evoked],mixed_norm->[_reapply_source_weighting,_compute_residual,_prepare_gain,_make_sparse_stc],_prepare_gain->[_prepare_gain_column],_prepare_gain_column->[_prepare_weights]]
Internal function to prepare the gain info whitener source_weighting mask for a single.
You can't go by the length, you need to look at the type. Discrete source spaces can be constructed with legitimate normals, for example.
@@ -869,6 +869,8 @@ dtx_17(void **state) int rc; int i; + skip(); + /* Assume I am the leader. */ for (i = 0; i < 10; i++) { struct dtx_handle *dth = NULL;
[void->[vos_fetch_begin,vts_dtx_prep_update,set_iov,d_iov_set,vos_hdl2cont,vts_dtx_commit_visibility,vos_dtx_abort,vos_dtx_check,vts_dtx_begin,daos_fail_loc_set,io_test_obj_update,vts_dtx_cos,vts_dtx_end,dts_buf_render,vos_dtx_commit,io_test_obj_fetch,malloc,vts_dtx_abort_visibility,daos_dti_equal,vts_dtx_shares,vos_dtx_aggregate,vos_dtx_del_cos,assert_int_equal,vos_dtx_list_cos,vos_dtx_lookup_cos,daos_dti_gen,vos_obj_punch,D_FREE,vos_dtx_add_cos,vos_iterate,sleep,vts_dtx_shares_with_punch,assert_memory_not_equal,daos_fail_loc_reset,vos_dtx_fetch_committable,crt_hlc_get,assert_memory_equal,free,vos_dtx_stat,d_hash_murmur64,D_FREE_PTR,memset,hash_key,vts_key_gen,assert_true,lrand48],run_dtx_tests->[cmocka_run_group_tests_name],int->[memcmp,assert_true,D_ASSERT,D_INIT_LIST_HEAD,test_args_reset,D_ALLOC_PTR]]
- - - - - - - - - - - - - - - - - - Method to update object in D - Bus. VOS_ITER_DKEY vts_dtx_iter_cb vts.
Why skip list dkey test?
@@ -30,10 +30,6 @@ import java.util.stream.Stream; import javax.inject.Inject; -import com.cloud.resource.icon.dao.ResourceIconDao; -import com.cloud.server.ResourceManagerUtil; -import com.cloud.storage.dao.VMTemplateDetailsDao; -import com.cloud.vm.VirtualMachineManager; import org.apache.cloudstack.acl.ControlledEntity.ACLType; import org.apache.cloudstack.affinity.AffinityGroupDomainMapVO; import org.apache.cloudstack.affinity.AffinityGroupResponse;
[QueryManagerImpl->[searchForIsosInternal->[searchForTemplatesInternal],searchForServiceOfferingsInternal->[getMinimumCpuServiceOfferingJoinSearchCriteria,getMinimumCpuSpeedServiceOfferingJoinSearchCriteria,findRelatedDomainIds,getMinimumMemoryServiceOfferingJoinSearchCriteria],buildAffinityGroupSearchCriteria->[buildAffinityGroupViewSearchCriteria,buildAffinityGroupViewSearchBuilder],searchForTemplatesInternal->[searchForTemplatesInternal],listResourceIcons->[listResourceIcons]]]
Reads a single object from the System. Imports all the command objects that are part of the API.
Is this reordering needed?
@@ -252,11 +252,16 @@ module Engine end def float_corporation(corporation) - # Designer says that bankrupt corps keep insolvency flag - clear_bankrupt(corporation) + clear_bankrupt!(corporation) super end + def action_processed(_action) + @corporations.each do |corporation| + make_bankrupt!(corporation) if corporation.share_price&.type == :close + end + end + def sorted_corporations @corporations.sort_by { |c| corp_layer(c) } end
[G1860->[par_prices->[corp_lo_par,par_prices,corp_hi_par,bankrupt?],float_corporation->[clear_bankrupt],path_distance_walk->[lane_match?,path_distance_walk],can_ipo?->[corp_layer],compute_distance_graph->[node_distance_walk,get_token_cities],sorted_corporations->[corp_layer],corporation_available?->[can_ipo?],clear_insolvent->[insolvent?],max_halts->[ignore_halts?],check_intersection->[visit_route],check_other->[check_hex_reentry],status_str->[insolvent?,bankrupt?],check_connected->[tokened_out?],route_trains->[insolvent?],make_insolvent->[insolvent?],compute_stops->[ignore_halt_subsidies?,ignore_halts?,ignore_second_allowance?],check_home_token->[get_token_cities],visit_route->[visit_route],node_distance_walk->[node_distance_walk]]]
Returns the float - valued corp object if it exists in the corpus.
you need to do this on every single action? that's gonna make the action * N
@@ -627,7 +627,7 @@ export class AmpForm { const name = success ? FormState_.SUBMIT_SUCCESS : FormState_.SUBMIT_ERROR; const event = createCustomEvent(this.win_, `${TAG}.${name}`, {response: json}); - this.actions_.trigger(this.form_, name, event); + this.actions_.trigger(this.form_, name, event, ActionTrust.MEDIUM); } /**
[No CFG could be retrieved]
Private helper methods for handling redirect responses. Retrieves the form data as an object.
Shouldn't these have high trust, too? Or at least their source trust.
@@ -38,7 +38,7 @@ namespace System.Diagnostics.Tracing #endif #if FEATURE_PERFTRACING - private readonly TraceLoggingEventHandleTable m_eventHandleTable = new TraceLoggingEventHandleTable(); + private readonly TraceLoggingEventHandleTable m_eventHandleTable = null!; #endif /// <summary>
[No CFG could be retrieved]
Creates an EventSource object with a given name for non - contract based events. Construct a new EventSource with a given name and a given config.
I don't understand this change. Where is it initialized now?
@@ -73,6 +73,7 @@ mkdir -p {envs}/test1/bin mkdir -p {envs}/test2/bin """ +@unittest.skipIf(skip_tests, "conda-env not installed on path") def test_activate_test1(): for shell in shells: with TemporaryDirectory(prefix='envs', dir=dirname(__file__)) as envs:
[test_CONDA_DEFAULT_ENV->[_write_entry_points],test_activate_test1_test2->[_write_entry_points],test_activate_test1_test3->[_write_entry_points],test_PS1_no_changeps1->[_write_entry_points],test_deactivate->[_write_entry_points],test_activate_test1_deactivate->[_write_entry_points],test_PS1->[_write_entry_points],test_activate_test3->[_write_entry_points],test_wrong_args->[_write_entry_points],test_activate_help->[_write_entry_points],test_activate_symlinking->[_write_entry_points],test_activate_test1->[_write_entry_points]]
Test that the test1 environment is activated.
Actually, I'd rather they failed. I don't see why conda-env wouldn't be installed as it's a dependency of conda, and if these tests are silently skipped, then we won't know if something in activate breaks.
@@ -652,6 +652,17 @@ public class ParDo { return withSideInputs(Arrays.asList(sideInputs)); } + /** + * Returns a new {@link ParDo} {@link PTransform} that's like this {@link PTransform} but with + * the specified additional side inputs. Does not modify this {@link PTransform}. + * + * <p>See the discussion of Side Inputs above for more explanation. + */ + public SingleOutput<InputT, OutputT> withSideInput(String tagId, PCollectionView<?> sideInput) { + sideInput.setTagInternalId(tagId); + return withSideInputs(sideInput); + } + /** * Returns a new {@link ParDo} {@link PTransform} that's like this {@link PTransform} but with * the specified additional side inputs. Does not modify this {@link PTransform}.
[ParDo->[codersForStateSpecTypes->[of],getDoFnSchemaInformation->[getFieldAccessDescriptorFromParameter],isSplittable->[isSplittable],MultiOutput->[expand->[validateStateApplicableForInput,validateWindowType,finishSpecifyingStateSpecs],populateDisplayData->[populateDisplayData],withSideInputs->[withSideInputs],toString->[toString],getKindString->[getFn]],SingleOutput->[populateDisplayData->[populateDisplayData],expand->[finishSpecifyingStateSpecs],toString->[toString],withSideInputs->[withSideInputs]]]]
Add side inputs to the output.
We shouldn't modify the input view. Rather clone it and modify the the cloned copy.
@@ -85,16 +85,4 @@ public interface Checkpointable { * restart from the beginning. */ UniqueId loadCheckpoint(ActorId actorId, List<Checkpoint> availableCheckpoints); - - /** - * Delete an expired checkpoint; - * - * This method will be called when an checkpoint is expired. You should implement this method to - * delete your application checkpoint data. Note, the maximum number of checkpoints kept in the - * backend can be configured at `RayConfig.num_actor_checkpoints_to_keep`. - * - * @param actorId ID of the actor. - * @param checkpointId ID of the checkpoint that has expired. - */ - void checkpointExpired(ActorId actorId, UniqueId checkpointId); }
[No CFG could be retrieved]
Load checkpoint from checkpoint cache.
@chaokunyang this whole class can be removed.
@@ -70,4 +70,10 @@ class ApplicationController < ActionController::Base def touch_current_user current_user.touch end + + private + + def core_pages? + false + end end
[ApplicationController->[customize_params->[to_s],after_sign_in_path_for->[ago,env,stored_location_for,created_at],not_found->[raise,new],valid_request_origin?->[to_s,origin,nil?,test?,raise,info,present?,gsub,start_with?],require_http_auth->[authenticate_or_request_with_http_basic],efficient_current_user_id->[present?,flatten],touch_current_user->[touch],raise_banned->[raise,banned],authenticate_user!->[redirect_to,render,respond_to,html,json],set_no_cache_header->[headers],helper_method,include,protect_from_forgery]]
Touch the current user s if it exists.
@rhymes this method common for alll controllers
@@ -365,7 +365,9 @@ public final class SystemSessionProperties false, value -> { boolean spillEnabled = (Boolean) value; - if (spillEnabled && featuresConfig.getSpillerSpillPaths().isEmpty()) { + if (spillEnabled + && featuresConfig.getSpillerSpillPaths().isEmpty() + && (!serverConfig.isCoordinator() || nodeSchedulerConfig.isIncludeCoordinator())) { throw new PrestoException( INVALID_SESSION_PROPERTY, format("%s cannot be set to true; no spill paths configured", SPILL_ENABLED));
[SystemSessionProperties->[isUnwrapCasts->[getSystemProperty],isDictionaryAggregationEnabled->[getSystemProperty],isOptimizeTopNRowNumber->[getSystemProperty],getFilterAndProjectMinOutputPageRowCount->[getSystemProperty],isPushTableWriteThroughUnion->[getSystemProperty],getQueryMaxCpuTime->[getSystemProperty],getRequiredWorkersMaxWait->[getSystemProperty],isEnableIntermediateAggregations->[getSystemProperty],isPushAggregationThroughOuterJoin->[getSystemProperty],isSpillOrderBy->[getSystemProperty],isLegacyTimestamp->[getSystemProperty],isSpatialJoinEnabled->[getSystemProperty],isUsePreferredWritePartitioning->[getSystemProperty],resourceOvercommit->[getSystemProperty],isCollectPlanStatisticsForAllQueries->[getSystemProperty],isLateMaterializationEnabled->[getSystemProperty],getInitialSplitsPerNode->[getSystemProperty],isGroupedExecutionEnabled->[getSystemProperty],isDistributedSortEnabled->[getSystemProperty],getFilterAndProjectMinOutputPageSize->[getSystemProperty],isSpillEnabled->[getSystemProperty],getMaxGroupingSets->[getSystemProperty],getQueryMaxRunTime->[getSystemProperty],isRedistributeWrites->[getSystemProperty],isPushPartialAggregationThroughJoin->[getSystemProperty],getQueryMaxTotalMemoryPerNode->[getSystemProperty],getRequiredWorkers->[getSystemProperty],preferPartialAggregation->[getSystemProperty],validateIntegerValue->[PrestoException,intValue,format],getDynamicFilteringMaxPerDriverSize->[getSystemProperty],useMarkDistinct->[getSystemProperty],isDynamicScheduleForGroupedExecution->[getSystemProperty],validateValueIsPowerOfTwo->[PrestoException,intValue,bitCount,format],planWithTableNodePartitioning->[getSystemProperty],isDistributedIndexJoinEnabled->[getSystemProperty],getQueryMaxStageCount->[getSystemProperty],isEnableForcedExchangeBelowGroupId->[getSystemProperty],isSkipRedundantSort->[getSystemProperty],getQueryMaxTotalMemory->[getSystemProperty],getHashPartitionCount->[getSystemProperty],getDynamicFilteringMaxPerDriverRowCount->[getSystemProperty],getQueryPriority->[checkArgument,getSystemProperty],getQueryMaxExecutionTime->[getSystemProperty],getJoinMaxBroadcastTableSize->[getSystemProperty],isFastInequalityJoin->[getSystemProperty],isEnableDynamicFiltering->[getSystemProperty],getExecutionPolicy->[getSystemProperty],preferStreamingOperators->[getSystemProperty],getConcurrentLifespansPerNode->[checkArgument,empty,getSystemProperty,of],isOptimizeDistinctAggregationEnabled->[getSystemProperty],validateNullablePositiveIntegerValue->[validateIntegerValue],isForceSingleNodeOutput->[getSystemProperty],isScaleWriters->[getSystemProperty],isNewOptimizerEnabled->[getSystemProperty],getOptimizerTimeout->[getSystemProperty],isPredicatePushdownUseTableProperties->[getSystemProperty],isShareIndexLoading->[getSystemProperty],getQueryMaxMemoryPerNode->[getSystemProperty],isStatisticsCpuTimerEnabled->[getSystemProperty],getTaskWriterCount->[getSystemProperty],getTaskConcurrency->[getSystemProperty],getWriterMinSize->[getSystemProperty],ignoreDownStreamPreferences->[getSystemProperty],getQueryMaxMemory->[getSystemProperty],getAggregationOperatorUnspillMemoryLimit->[checkArgument,getSystemProperty,toBytes],getJoinReorderingStrategy->[getSystemProperty],isOptimizeMetadataQueries->[getSystemProperty],isOptimizeHashGenerationEnabled->[getSystemProperty],getSpatialPartitioningTableName->[getSystemProperty,ofNullable],isSpillWindowOperator->[getSystemProperty],isDefaultFilterFactorEnabled->[getSystemProperty],getMaxReorderedJoins->[getSystemProperty],getJoinDistributionType->[getSystemProperty],isEnableStatsCalculator->[getSystemProperty],isIgnoreStatsCalculatorFailures->[getSystemProperty],isParseDecimalLiteralsAsDouble->[getSystemProperty],isColocatedJoinEnabled->[getSystemProperty],isExchangeCompressionEnabled->[getSystemProperty],getMaxDriversPerTask->[empty,getSystemProperty,of],getSplitConcurrencyAdjustmentInterval->[getSystemProperty],isUnwrapCasts,isOptimizeTopNRowNumber,getFilterAndProjectMinOutputPageRowCount,isPushTableWriteThroughUnion,getQueryMaxCpuTime,isEmpty,intValue,getRequiredWorkersMaxWait,isEnableIntermediateAggregations,isOptimizeHashGeneration,isPushAggregationThroughOuterJoin,NodeMemoryConfig,format,isSpillOrderBy,isLegacyTimestamp,isPushPartialAggregationThoughJoin,isDynamicScheduleForGroupedExecutionEnabled,isUsePreferredWritePartitioning,isDistributedIndexJoinsEnabled,isCollectPlanStatisticsForAllQueries,PrestoException,isLateMaterializationEnabled,getInitialSplitsPerNode,isGroupedExecutionEnabled,isPreferPartialAggregation,isDistributedSortEnabled,getFilterAndProjectMinOutputPageSize,stringProperty,isSpillEnabled,getMaxGroupingSets,MemoryManagerConfig,getQueryMaxRunTime,booleanProperty,isRedistributeWrites,getRequiredWorkers,durationProperty,getWriterCount,getMaxQueryMemory,getDynamicFilteringMaxPerDriverSize,getMaxQueryTotalMemory,validateValueIsPowerOfTwo,getIterativeOptimizerTimeout,TaskManagerConfig,isIgnoreDownstreamPreferences,getMaxStageCount,getInitialHashPartitions,isEnableForcedExchangeBelowGroupId,isSkipRedundantSort,FeaturesConfig,min,getDynamicFilteringMaxPerDriverRowCount,integerProperty,getQueryMaxExecutionTime,getMaxQueryMemoryPerNode,getJoinMaxBroadcastTableSize,isEnableDynamicFiltering,getMaxQueryTotalMemoryPerNode,getQueryExecutionPolicy,isColocatedJoinsEnabled,getConcurrentLifespansPerTask,QueryManagerConfig,validateNullablePositiveIntegerValue,isOptimizeMixedDistinctAggregations,isForceSingleNodeOutput,isScaleWriters,enumProperty,isPredicatePushdownUseTableProperties,isShareIndexLoading,isStatisticsCpuTimerEnabled,getTaskConcurrency,getWriterMinSize,getAggregationOperatorUnspillMemoryLimit,getJoinReorderingStrategy,dataSizeProperty,isOptimizeMetadataQueries,of,isSpillWindowOperator,isSpatialJoinsEnabled,isUseMarkDistinct,isDefaultFilterFactorEnabled,getMaxReorderedJoins,getJoinDistributionType,isEnableStatsCalculator,isIgnoreStatsCalculatorFailures,isFastInequalityJoins,isParseDecimalLiteralsAsDouble,isIterativeOptimizerEnabled,isExchangeCompressionEnabled,isDictionaryAggregation,getMaxDriversPerTask,getSplitConcurrencyAdjustmentInterval]]
Protected boolean property. This function is called by the memory manager when memory is not available.
this is correct, complex, weak (coordinator is unlikely included in any reasonable setup) and insufficient (doesnt check whether workers have spill paths configured) i am fine with this, but i would also be fine with this check being removed altogether
@@ -314,13 +314,14 @@ def filter_products_by_stock_availability(qs, stock_availability, channel_slug): ) allocated_subquery = Subquery(queryset=allocations, output_field=IntegerField()) - stocks = list( + stocks = ( Stock.objects.for_channel(channel_slug) .filter(quantity__gt=Coalesce(allocated_subquery, 0)) - .values_list("product_variant_id", flat=True) + .values("product_variant_id") ) - - variants = ProductVariant.objects.filter(pk__in=stocks).values("product_id") + variants = ProductVariant.objects.filter( + Exists(stocks.filter(product_variant_id=OuterRef("pk"))) + ).values("product_id") if stock_availability == StockAvailability.IN_STOCK: qs = qs.filter(Exists(variants.filter(product_id=OuterRef("pk"))))
[_filter_minimal_price->[filter_products_by_minimal_price],_filter_stock_availability->[filter_products_by_stock_availability],_filter_variant_price->[filter_products_by_variant_price],filter_categories->[filter_products_by_categories],_filter_attributes->[filter_products_by_attributes],ProductFilter->[filter_stock_availability->[_filter_stock_availability],filter_is_published->[_filter_products_is_published],filter_minimal_price->[_filter_minimal_price],filter_attributes->[_filter_attributes],filter_variant_price->[_filter_variant_price]],filter_products_by_attributes->[_clean_product_attributes_boolean_filter_input,filter_products_by_attributes_values,_clean_product_attributes_date_time_range_filter_input,_clean_product_attributes_range_filter_input,_clean_product_attributes_filter_input],CollectionFilter->[filter_is_published->[_filter_collections_is_published]],filter_search->[product_search],filter_collections->[filter_products_by_collections]]
Filter products by stock availability.
aren't brackets here unnecessary?
@@ -856,6 +856,7 @@ module.exports = JhipsterServerGenerator.extend({ // Create Gatling test files if (this.testFrameworks.indexOf('gatling') !== -1) { this.copy(TEST_DIR + 'gatling/conf/gatling.conf', TEST_DIR + 'gatling/conf/gatling.conf'); + this.copy(TEST_DIR = 'gatling/conf/logback.xml', TEST_DIR + 'gatling/conf/logback.xml'); mkdirp(TEST_DIR + 'gatling/data'); mkdirp(TEST_DIR + 'gatling/bodies'); mkdirp(TEST_DIR + 'gatling/simulations');
[No CFG could be retrieved]
Create the test files for the given application type. Create Cucumber test files.
typo here, it should be `+` instead of `=`
@@ -79,11 +79,12 @@ def _load_weights( model: QuantizableMobileNetV3, model_url: Optional[str], progress: bool, + strict: bool ) -> None: if model_url is None: raise ValueError("No checkpoint is available for {}".format(arch)) state_dict = load_state_dict_from_url(model_url, progress=progress) - model.load_state_dict(state_dict) + model.load_state_dict(state_dict, strict=strict) def _mobilenet_v3_model(
[_mobilenet_v3_model->[QuantizableMobileNetV3,_load_weights,fuse_model],QuantizableMobileNetV3->[fuse_model->[fuse_model]],mobilenet_v3_large->[_mobilenet_v3_model]]
Load the weights of a single critical block.
nit: default to True, which is desired in most of cases.
@@ -169,10 +169,10 @@ func (rs *RegistrySynchronizer) handleUpkeepPerformedLogs(done func()) { func (rs *RegistrySynchronizer) handleUpkeepPerformed(broadcast log.Broadcast) { txHash := broadcast.RawLog().TxHash.Hex() - logger.Debugw("RegistrySynchronizer: processing UpkeepPerformed log", "jobID", rs.job.ID, "txHash", txHash) + rs.logger.Debugw("RegistrySynchronizer: processing UpkeepPerformed log", "txHash", txHash) was, err := rs.logBroadcaster.WasAlreadyConsumed(rs.orm.DB, broadcast) if err != nil { - logger.Warn(errors.Wrapf(err, "RegistrySynchronizer: unable to check if log was consumed, jobID: %d", rs.job.ID)) + rs.logger.WithError(err).Warn("RegistrySynchronizer: unable to check if log was consumed") return } if was {
[HandleUpkeepRegistered->[Wrapf,ErrorIf,Error,Warn,Hex,DecodedLog,syncUpkeep,Debugw,DefaultQueryCtx,WithContext,String,Errorf,MarkConsumed,Int64,WasAlreadyConsumed,RawLog],handleUpkeepPerformedLogs->[Errorf,Retrieve,handleUpkeepPerformed],handleUpkeepRegisteredLogs->[HandleUpkeepRegistered,Wrapf,Error,DefaultQueryCtx,Errorf,RegistryForJob,Retrieve],handleUpkeepCancelled->[BatchDeleteUpkeepsForJob,Wrapf,ErrorIf,Error,Warn,Hex,DecodedLog,Sprintf,Debugw,DefaultQueryCtx,WithContext,String,Errorf,MarkConsumed,Int64,WasAlreadyConsumed,RawLog],processLogs->[handleUpkeepRegisteredLogs,handleUpkeepCanceledLogs,handleSyncRegistryLog,Wait,Add,handleUpkeepPerformedLogs],handleUpkeepPerformed->[Wrapf,ErrorIf,Error,Warn,Hex,DecodedLog,SetLastRunHeightForUpkeepOnJob,Debugw,DefaultQueryCtx,WithContext,String,Errorf,MarkConsumed,Int64,WasAlreadyConsumed,RawLog],handleSyncRegistryLog->[Wrapf,ErrorIf,Error,Warn,Hex,syncRegistry,Debugw,DefaultQueryCtx,WithContext,Retrieve,String,Errorf,MarkConsumed,WasAlreadyConsumed,RawLog],handleUpkeepCanceledLogs->[handleUpkeepCancelled,Errorf,Retrieve]]
handleUpkeepPerformed checks if the log was consumed and if so marks it as consumed.
We can probably remove the "RegistrySynchronizer: "... portion of these error messages now, since the named logger already specifies this is for the registry synchronizer.
@@ -54,7 +54,7 @@ namespace Dnn.ExportImport.Components.Interfaces IList<ExportModulePermission> GetModulePermissions(int moduleId, DateTime toDate, DateTime? fromDate); IList<ExportTabModule> GetTabModules(int tabId, bool includeDeleted, DateTime toDate, DateTime? fromDate); - IList<ExportTabModuleSetting> GetTabModuleSettings(int tabId, DateTime toDate, DateTime? fromDate); + IList<ExportTabModuleSetting> GetTabModuleSettings(int tabId, bool includeDeleted, DateTime toDate, DateTime? fromDate); PermissionInfo GetPermissionInfo(string permissionCode, string permissionKey, string permissionName);
[No CFG could be retrieved]
Get all tab modules.
This is a breaking change. We need to keep the old signature and add a new method to the interface
@@ -769,6 +769,11 @@ class ASTConverter(ast35.NodeTransformer): else: return UnicodeExpr(n.s) + # JoinedStr(expr* values) + @with_line + def visit_JoinedStr(self, n: ast35.JoinedStr) -> StrExpr: + return StrExpr(n.values[0]) + # Bytes(bytes s) @with_line def visit_Bytes(self, n: ast35.Bytes) -> Union[BytesExpr, StrExpr]:
[TypeConverter->[visit_Str->[parse_type_comment],visit_Subscript->[translate_expr_list],visit_raw_str->[parse_type_comment],visit_List->[translate_expr_list],visit_Tuple->[translate_expr_list]],parse->[parse],ASTConverter->[visit_Delete->[translate_expr_list],visit_AsyncWith->[as_block],visit_While->[as_block],visit_If->[as_block],do_func_def->[in_class,parse,translate_expr_list,as_block],visit_BinOp->[from_operator],visit_ClassDef->[stringify_name,find,translate_expr_list,as_block],visit_AugAssign->[from_operator],visit_Dict->[translate_expr_list],visit_BoolOp->[group->[group],group,translate_expr_list],visit_AsyncFor->[as_block],visit_With->[as_block],visit_Assign->[translate_expr_list,parse_type_comment],stringify_name->[stringify_name],visit_Import->[translate_module_id],transform_args->[make_argument],as_block->[translate_stmt_list],visit_Try->[as_block],visit_Lambda->[as_block,transform_args],visit_Compare->[from_comp_operator,translate_expr_list],visit_DictComp->[translate_expr_list],visit_ExtSlice->[translate_expr_list],visit_GeneratorExp->[translate_expr_list],visit_For->[as_block],visit_Set->[translate_expr_list],visit_Module->[fix_function_overloads,translate_stmt_list],visit_Call->[is_star2arg,translate_expr_list],visit_ImportFrom->[translate_module_id]],parse_type_comment->[parse]]
Visit a string node and return a .
This whole method should be inside something like `if hasattr(ast35, 'JoinedStr'):` so that mypy will still work with the previous version of typed_ast.
@@ -82,6 +82,11 @@ func showRegionCommandFunc(cmd *cobra.Command, args []string) { fmt.Printf("Failed to get region: %s\n", err) return } + if flag := cmd.Flag("jq"); flag != nil && flag.Value.String() != "" { + printWithJQFilter(r, flag.Value.String()) + return + } + fmt.Println(r) }
[Printf,Flags,NewBuffer,UsageString,Println,ReadByte,Next,Lookup,String,Sscanf,AddCommand,Atoi]
showRegionCommandFunc returns a Command object that displays the region_id and the region_ showRegionTopReadCommandFunc returns a region with key subcommand.
can we provide a Print interface for jq or standard output?
@@ -2023,6 +2023,9 @@ func bindDefaultValue(path string, value interface{}, spec *DefaultSpec, typ Typ for name, raw := range spec.Language { language[name] = json.RawMessage(raw) } + if spec.Environment == nil { + diags = diags.Append(errorf(path, "Default must specify an environment")) + } dv.Environment, dv.Language = spec.Environment, language }
[marshalResource->[marshalObjectData,marshalObject],bindProperties->[newOptionalType,bindType],bindResourceType->[bindResourceTypeDetails],MarshalYAML->[MarshalSpec],bindEnumTypeDetails->[bindPrimitiveType],marshalType->[marshalType],bindObjectType->[bindObjectTypeDetails],marshalFunction->[marshalObject],bindType->[newMapType,bindPrimitiveType,bindType,bindTypeSpecRef,newArrayType,newInputType,newUnionType],MarshalSpec->[String,IsInputShape],MarshalJSON->[MarshalSpec],bindTypeSpecRef->[GetType,bindPrimitiveType,GetResourceType,parseTypeSpecRef],bindEnumType->[bindEnumTypeDetails],marshalObject->[marshalObjectData],bindObjectTypeDetails->[bindProperties],String->[String],newUnionType->[String],bindProperties,bindEnumTypeDetails,bindObjectType,ImportLanguages,bindObjectTypeDetails,String]
BindSpec converts a serializable PackageSpec into a Package and then imports it into the internal package ilitates interning of the objects.
This is not correct: it is valid for defaults to provide only static values and no env vars. It seems reasonable to require one or the other, though.
@@ -79,7 +79,11 @@ namespace Microsoft.NET.HostModel.Tests fileSpecs.Add(new FileSpec(BundleHelper.GetAppPath(fixture), "app.repeat")); Bundler bundler = new Bundler(hostName, bundleDir.FullName); - Assert.Throws<ArgumentException>(() => bundler.GenerateBundle(fileSpecs)); + + var argEx = Assert.Throws<ArgumentException>(() => bundler.GenerateBundle(fileSpecs)); + Assert.Equal("Invalid input specification: Found multiple entries with the same BundleRelativePath:" + + Environment.NewLine + + fileSpecs[1], argEx.Message); } [InlineData(true)]
[BundlerConsistencyTests->[SharedTestState->[Dispose->[Dispose]]]]
Tests if a file specification with duplicate entries fails.
Shouldn't the message contain both `fileSpecs[1]` and `fileSpecs[2]`. I think this is why the test is failing.
@@ -956,6 +956,7 @@ func parseHunks(curFile *DiffFile, maxLines, maxLineCharacters int, input *bufio lastLeftIdx := -1 leftLine, rightLine := 1, 1 + engine := db.GetEngine(db.DefaultContext) for { for isFragment { curFile.IsIncomplete = true
[GetComputedInlineDiffFor->[GetLine],GetTailSection]
parseHunks parses the diff file and returns the parsed diff and the error. ReadLine reads a single line of a header from the input stream and returns it as an.
Really not a fan of letting the engine spill in to services - if we're gonna do this what's the point of the functions in models.
@@ -359,6 +359,17 @@ public class KsqlConfig extends AbstractConfig { private static final String KSQL_PROPERTIES_OVERRIDES_DENYLIST_DOC = "Comma-separated list of " + "properties that KSQL users cannot override."; + public static final String KSQL_TOTAL_CACHE_MAX_BYTES_BUFFERING = + "ksql.cache.max.bytes.buffering.total"; + public static final long KSQL_TOTAL_CACHE_MAX_BYTES_BUFFERING_DEFAULT = -1; + public static final String KSQL_TOTAL_CACHE_MAX_BYTES_BUFFERING_DOC = "Limit on the total bytes " + + "used by Kafka Streams cache across all persistent queries"; + public static final String KSQL_TOTAL_CACHE_MAX_BYTES_BUFFERING_TRANSIENT = + "ksql.transient.cache.max.bytes.buffering.total"; + public static final long KSQL_TOTAL_CACHE_MAX_BYTES_BUFFERING_TRANSIENT_DEFAULT = -1; + public static final String KSQL_TOTAL_CACHE_MAX_BYTES_BUFFERING_TRANSIENT_DOC = "Limit on the " + + "total bytes used by Kafka Streams cache across all transient queries"; + public static final String KSQL_VARIABLE_SUBSTITUTION_ENABLE = "ksql.variable.substitution.enable"; public static final boolean KSQL_VARIABLE_SUBSTITUTION_ENABLE_DEFAULT = true;
[KsqlConfig->[buildConfigDef->[defineLegacy,defineCurrent],ConfigValue->[isResolved->[isResolved]],getKsqlStreamConfigProps->[getKsqlStreamConfigProps],buildStreamingConfig->[applyStreamsConfig],streamTopicConfigNames->[getName],getAllConfigPropsWithSecretsObfuscated->[getKsqlConfigPropsWithSecretsObfuscated],overrideBreakingConfigsWithOriginalValues->[KsqlConfig],cloneWithPropertyOverwrite->[KsqlConfig,getKsqlStreamConfigProps,buildStreamingConfig],getKsqlStreamConfigPropsWithSecretsObfuscated->[convertToObfuscatedString],resolveStreamsConfig->[ConfigValue],CompatibilityBreakingConfigDef->[defineCurrent->[define],defineLegacy->[define],define->[define]],configDef,buildStreamingConfig]]
This class defines the parameters for the error messages. Creates CompatibilityBreakingConfigDef from a list of objects.
I think based on some of the other query configs we have already, these two configs name would be more suitable `ksql.query.transient.max.bytes.buffering.total` `ksql.query.persistent.max.bytes.buffering.total`
@@ -296,8 +296,6 @@ class TopicUser < ActiveRecord::Base FROM topic_users AS ftu WHERE ftu.user_id = :user_id and ftu.topic_id = :topic_id)" - INSERT_TOPIC_USER_SQL_STAFF = INSERT_TOPIC_USER_SQL.gsub("highest_post_number", "highest_staff_post_number") - def update_last_read(user, topic_id, post_number, new_posts_read, msecs, opts = {}) return if post_number.blank? msecs = 0 if msecs.to_i < 0
[TopicUser->[update_last_read->[update_last_read,notification_level_change],create_missing_record->[notification_levels,notification_reasons],ensure_consistency!->[update_post_action_cache],track_visit!->[change],change->[notification_reasons]]]
Updates the last read of a topic post. if there is no record in DB insert it else raise a new record.
No longer required because the SQL query doesn't include `highest_post_number` anymore.
@@ -428,4 +428,13 @@ public final class HAProxyMessage { public int destinationPort() { return destinationPort; } + + /** + * Returns a list of {@link HAProxyTLV} or an empty list if no TLVs are present. + * <p> + * TLVs are only available for the Proxy Protocol V2 + */ + public List<HAProxyTLV> getTlvs() { + return Collections.unmodifiableList(tlvs); + } }
[HAProxyMessage->[decodeHeader->[HAProxyMessage],HAProxyMessage]]
Returns the destination port.
Why not store the `tlvs` already as immutable ?
@@ -138,6 +138,16 @@ std::string wide_to_utf8(const std::wstring &input) return out; } +void wide_add_codepoint(std::wstring &result, char32_t codepoint) +{ + if ((0xD800 <= codepoint && codepoint <= 0xDFFF) || (0x10FFFF < codepoint)) { + // Invalid codepoint, replace with unicode replacement character + result.push_back(0xFFFD); + return; + } + result.push_back(codepoint); +} + #else // _WIN32 std::wstring utf8_to_wide(const std::string &input)
[No CFG could be retrieved]
_WIN32 - > _WIN32 - > _UTF8Recursive _WIN32 ByByteToWideChar - convert a UTF - 8 string into a.
Couldn't the code for checking if the code point is valid be shared between the two implementations?
@@ -226,7 +226,7 @@ EOT } $settingKey = $input->getArgument('setting-key'); - if (!$settingKey) { + if (!$settingKey || !is_string($settingKey)) { return 0; }
[ConfigCommand->[listConfiguration->[listConfiguration]]]
Executes the command. Get all configuration values. Returns an array of configuration values that can be used to configure a specific configuration. Returns a list of possible configuration values for a specific configuration.
This adds no value AFAICT.
@@ -216,7 +216,7 @@ func (c *baseClient) getMembers(ctx context.Context, url string) (*pdpb.GetMembe members, err := pdpb.NewPDClient(cc).GetMembers(ctx, &pdpb.GetMembersRequest{}) if err != nil { attachErr := errors.Errorf("error:%s target:%s status:%s", err, cc.Target(), cc.GetState().String()) - return nil, errors.WithStack(attachErr) + return nil, errs.ErrClientGetMember.GenWithStackByArgs(attachErr) } return members, nil }
[updateURLs->[Strings,Info,GetClientUrls,DeepEqual],updateLeader->[updateURLs,Warn,WithTimeout,switchLeader,ZapError,WithStack,GetMembers,String,Errorf,GetClientUrls,getMembers,Done,GetLeader],getOrCreateGRPCConn->[GetClientConn,Unlock,WithTimeout,Target,RLock,Close,WithStack,ToTLSConfig,Lock,String,RUnlock,GetState,Debug],initClusterID->[Warn,WithTimeout,WithCancel,ZapError,WithStack,String,getMembers,GetClusterId,GetHeader],leaderLoop->[updateLeader,Error,WithCancel,ZapError,After,Done],switchLeader->[getOrCreateGRPCConn,Info,Unlock,RLock,String,Lock,RUnlock],GetLeaderAddr->[RUnlock,RLock],getMembers->[getOrCreateGRPCConn,Target,WithStack,NewPDClient,GetMembers,Errorf,String,GetState],initRetry->[WithStack,After,Done],leaderLoop,Uint64,WithCancel,cancel,Info,initRetry,Add]
getMembers returns a list of members.
Should we change it to `errs.ErrClientGetMember.Wrap(errors.WithStack(attachErr)).FastGenWithCause()`?
@@ -2080,7 +2080,9 @@ def _add_common_install_arguments(parser, build_help, lockfile=True): parser.add_argument("-r", "--remote", action=OnceArgument, help='Look in the specified remote server') parser.add_argument("-u", "--update", action='store_true', default=False, - help="Check updates exist from upstream remotes") + help="Check updates exist for the current reference from upstream remotes." + " It doesn't check for updates for the dependencies, clean the local" + " cache to update also dependencies.") if lockfile: parser.add_argument("-l", "--lockfile", action=OnceArgument, nargs='?', const=".", help="Path to a lockfile or folder containing 'conan.lock' file. "
[Command->[export->[export],info->[info],install->[install],editable->[info],source->[source],remove->[remove,info],new->[new],_print_similar->[_commands],imports->[imports],upload->[upload],copy->[copy],download->[download],run->[_warn_python_version,_print_similar,_commands,_show_help],export_pkg->[export_pkg],test->[test],inspect->[inspect],package->[package],_show_help->[check_all_commands_listed],_warn_python_version->[get],create->[create],build->[build]],main->[Command,run]]
Add common install arguments to a parser.
but --update checks for updates in dependencies too.
@@ -48,6 +48,17 @@ def clip_by_value(t, clip_value_min, clip_value_max, Note: `clip_value_min` needs to be smaller or equal to `clip_value_max` for correct results. + For example: + + ```python + A=tf.constant([[1,20,13],[3,21,13]]) + B=tf.clip_by_value(A, clip_value_min=0, clip_value_max=3) #[[1, 3, 3],[3, 3, 3]] + C=tf.clip_by_value(A, clip_value_min=0., clip_value_max=3.) + # throws `TypeError` as input and clip_values are of different dtype + D=tf.constant([2.0,3.1,5.3]) + E=tf.clip_by_value(D, clip_value_min=1., clip_value_max=3.) #[2.,3.,3.] + ``` + Args: t: A `Tensor` or `IndexedSlices`. clip_value_min: A 0-D (scalar) `Tensor`, or a `Tensor` with the same shape
[clip_by_value->[minimum,IndexedSlices,maximum,merge_with,isinstance,name_scope,convert_to_tensor],global_norm->[TypeError,list,colocate_with,reduce_sum,append,l2_loss,enumerate,sqrt,stack,isinstance,constant,name_scope,convert_to_tensor],clip_by_global_norm->[TypeError,list,minimum,colocate_with,append,global_norm,enumerate,is_finite,IndexedSlices,zip,identity,where,isinstance,constant,name_scope,float,convert_to_tensor],clip_by_average_norm->[rsqrt,reduce_sum,rank,minimum,size,cast,identity,name_scope,range,convert_to_tensor,constant],_clip_by_value_grad->[broadcast_gradient_args,reduce_sum,logical_or,less,greater,where,zeros,reshape,shape],clip_by_norm->[ones_like,reduce_sum,sqrt,maximum,IndexedSlices,where,merge_with,isinstance,identity,name_scope,convert_to_tensor],tf_export,deprecated_endpoints,deprecated]
Clip tensor values to a specified min and max and return a new tensor with the same shape t_max - > IndexSlices t_min - > IndexSlices t_max.
Can the code be updated to be compatible with TF style? This means space around = used for assignment, space after commas, etc
@@ -380,8 +380,7 @@ namespace System.Dynamic.Tests return testObjects.SelectMany(i => testObjects.Select(j => new[] { i, j })); } - // We're not testing compilation, but we do need Reflection.Emit for the test - [ConditionalTheory(typeof(PlatformDetection), nameof(PlatformDetection.IsNotLinqExpressionsBuiltWithIsInterpretingOnly))] + [Theory] [MemberData(nameof(SameNameObjectPairs))] public void OperationOnTwoObjectsDifferentTypesOfSameName(object x, object y) {
[InvokeMemberBindingTests->[GenericMethod->[TellType],ByRef->[TryParseInt],MethodHiding->[Method],NoArgumentMatch->[Method],ManyArities->[GetValue,GetValue2]]]
Method that returns objects of the same name in the same assembly.
Also don't see anything ref.emitting. Was probably some .NET Native issue related to dynamic.
@@ -0,0 +1,4 @@ +from .OrderEvent import OrderEvent + +__all__ = [ + "OrderEvent"]
[No CFG could be retrieved]
No Summary Found.
Is there really a need for separate files for each model? I think we won't have that many of them. Let's think of such improvements when there is really an identified problem. Also, the `OrderEvents.py` filename doesn't follow the naming convention for Python modules that we use.
@@ -52,16 +52,15 @@ class TextAnalyticsClientOperationsMixin(object): api_version = self._get_api_version('entities_linking') if api_version == 'v3.0': from .v3_0.operations import TextAnalyticsClientOperationsMixin as OperationClass - elif api_version == 'v3.1-preview.1': - from .v3_1_preview_1.operations import TextAnalyticsClientOperationsMixin as OperationClass elif api_version == 'v3.1-preview.2': from .v3_1_preview_2.operations import TextAnalyticsClientOperationsMixin as OperationClass else: - raise NotImplementedError("APIVersion {} is not available".format(api_version)) + raise ValueError("API version {} does not have operation 'entities_linking'".format(api_version)) mixin_instance = OperationClass() mixin_instance._client = self._client mixin_instance._config = self._config mixin_instance._serialize = Serializer(self._models_dict(api_version)) + mixin_instance._serialize.client_side_validation = False mixin_instance._deserialize = Deserializer(self._models_dict(api_version)) return mixin_instance.entities_linking(documents, model_version, show_stats, **kwargs)
[TextAnalyticsClientOperationsMixin->[key_phrases->[key_phrases],sentiment->[sentiment],entities_linking->[entities_linking],entities_recognition_general->[entities_recognition_general],entities_recognition_pii->[entities_recognition_pii],languages->[languages]]]
Links entities from a well - known knowledge base. A base class for all Entities in a given set of documents. A base class for all entities recognition.
in the old regeneration, I deleted the v3.1-preview.1 generated folder, but didn't regenerate without that API version. This was also a bug before.