idx int64 0 165k | question stringlengths 73 4.15k | target stringlengths 5 918 | len_question int64 21 890 | len_target int64 3 255 |
|---|---|---|---|---|
155,400 | public void writeAll ( java . sql . ResultSet rs , boolean includeColumnNames ) throws SQLException , IOException { if ( includeColumnNames ) { writeColumnNames ( rs ) ; } while ( rs . next ( ) ) { writeNext ( resultService . getColumnValues ( rs ) ) ; } } | Writes the entire ResultSet to a CSV file . | 67 | 11 |
155,401 | public void addRecord ( String key , String value ) throws TarMalformatException , IOException { if ( key == null || value == null || key . length ( ) < 1 || value . length ( ) < 1 ) { throw new TarMalformatException ( RB . singleton . getString ( RB . ZERO_WRITE ) ) ; } int lenWithoutIlen = key . length ( ) + value . length ( ) + 3 ; // "Ilen" means Initial Length field. +3 = SPACE + = + \n int lenW = 0 ; // lenW = Length With initial-length-field if ( lenWithoutIlen < 8 ) { lenW = lenWithoutIlen + 1 ; // Takes just 1 char to report total } else if ( lenWithoutIlen < 97 ) { lenW = lenWithoutIlen + 2 ; // Takes 2 chars to report this total } else if ( lenWithoutIlen < 996 ) { lenW = lenWithoutIlen + 3 ; // Takes 3... } else if ( lenWithoutIlen < 9995 ) { lenW = lenWithoutIlen + 4 ; // ditto } else if ( lenWithoutIlen < 99994 ) { lenW = lenWithoutIlen + 5 ; } else { throw new TarMalformatException ( RB . singleton . getString ( RB . PIF_TOOBIG , 99991 ) ) ; } writer . write ( Integer . toString ( lenW ) ) ; writer . write ( ' ' ) ; writer . write ( key ) ; writer . write ( ' ' ) ; writer . write ( value ) ; writer . write ( ' ' ) ; writer . flush ( ) ; // Does this do anything with a BAOS? } | I guess the initial length field is supposed to be in units of characters not bytes? | 366 | 17 |
155,402 | synchronized public void shutdown ( ) { m_shutdown . set ( true ) ; if ( m_isExecutorServiceLocal ) { try { m_es . shutdown ( ) ; m_es . awaitTermination ( 365 , TimeUnit . DAYS ) ; } catch ( InterruptedException e ) { repairLog . warn ( "Unexpected interrupted exception" , e ) ; } } } | shutdown silences the babysitter and causes watches to not reset . Note that shutting down will churn ephemeral ZK nodes - shutdown allows the programmer to not set watches on nodes from terminated session . | 84 | 41 |
155,403 | public static Pair < BabySitter , List < String > > blockingFactory ( ZooKeeper zk , String dir , Callback cb ) throws InterruptedException , ExecutionException { ExecutorService es = CoreUtils . getCachedSingleThreadExecutor ( "Babysitter-" + dir , 15000 ) ; Pair < BabySitter , List < String > > babySitter = blockingFactory ( zk , dir , cb , es ) ; babySitter . getFirst ( ) . m_isExecutorServiceLocal = true ; return babySitter ; } | Create a new BabySitter and block on reading the initial children list . | 122 | 15 |
155,404 | public static Pair < BabySitter , List < String > > blockingFactory ( ZooKeeper zk , String dir , Callback cb , ExecutorService es ) throws InterruptedException , ExecutionException { BabySitter bs = new BabySitter ( zk , dir , cb , es ) ; List < String > initialChildren ; try { initialChildren = bs . m_eventHandler . call ( ) ; } catch ( Exception e ) { throw new ExecutionException ( e ) ; } return new Pair < BabySitter , List < String > > ( bs , initialChildren ) ; } | Create a new BabySitter and block on reading the initial children list . Use the provided ExecutorService to queue events to rather than creating a private ExecutorService . The initial set of children will be retrieved in the current thread and not the ExecutorService because it is assumed this is being called from the ExecutorService | 127 | 65 |
155,405 | public static BabySitter nonblockingFactory ( ZooKeeper zk , String dir , Callback cb , ExecutorService es ) throws InterruptedException , ExecutionException { BabySitter bs = new BabySitter ( zk , dir , cb , es ) ; bs . m_es . submit ( bs . m_eventHandler ) ; return bs ; } | Create a new BabySitter and make sure it reads the initial children list . Use the provided ExecutorService to queue events to rather than creating a private ExecutorService . | 81 | 35 |
155,406 | void checkClosed ( ) throws SQLException { if ( isClosed ) { throw Util . sqlException ( ErrorCode . X_07501 ) ; } if ( connection . isClosed ) { close ( ) ; throw Util . sqlException ( ErrorCode . X_08503 ) ; } } | An internal check for closed statements . | 67 | 7 |
155,407 | void performPostExecute ( ) throws SQLException { resultOut . clearLobResults ( ) ; generatedResult = null ; if ( resultIn == null ) { return ; } Result current = resultIn ; while ( current . getChainedResult ( ) != null ) { current = current . getUnlinkChainedResult ( ) ; if ( current . getType ( ) == ResultConstants . WARNING ) { SQLWarning w = Util . sqlWarning ( current ) ; if ( rootWarning == null ) { rootWarning = w ; } else { rootWarning . setNextWarning ( w ) ; } } else if ( current . getType ( ) == ResultConstants . ERROR ) { errorResult = current ; } else if ( current . getType ( ) == ResultConstants . DATA ) { generatedResult = current ; } } if ( resultIn . isData ( ) ) { currentResultSet = new JDBCResultSet ( connection . sessionProxy , this , resultIn , resultIn . metaData , connection . connProperties ) ; } } | processes chained warnings and any generated columns result set | 222 | 10 |
155,408 | boolean getMoreResults ( int current ) throws SQLException { checkClosed ( ) ; if ( resultIn == null || ! resultIn . isData ( ) ) { return false ; } if ( resultSetCounter == 0 ) { resultSetCounter ++ ; return true ; } if ( currentResultSet != null && current != KEEP_CURRENT_RESULT ) { currentResultSet . close ( ) ; } resultIn = null ; return false ; } | Note yet correct for multiple ResultSets . Should keep track of the previous ResultSet objects to be able to close them | 98 | 24 |
155,409 | void closeResultData ( ) throws SQLException { if ( currentResultSet != null ) { currentResultSet . close ( ) ; } if ( generatedResultSet != null ) { generatedResultSet . close ( ) ; } generatedResultSet = null ; generatedResult = null ; resultIn = null ; } | See comment for getMoreResults . | 65 | 7 |
155,410 | public boolean compatibleWithTable ( VoltTable table ) { String candidateName = getTableName ( table ) ; // table can't have the same name as the view if ( candidateName . equals ( viewName ) ) { return false ; } // view is for a different table if ( candidateName . equals ( srcTableName ) == false ) { return false ; } try { // ignore ret value here - just looking to not throw int groupColIndex = table . getColumnIndex ( groupColName ) ; VoltType groupColType = table . getColumnType ( groupColIndex ) ; if ( groupColType == VoltType . DECIMAL ) { // no longer a good type to group return false ; } // check the sum col is still value int sumColIndex = table . getColumnIndex ( sumColName ) ; VoltType sumColType = table . getColumnType ( sumColIndex ) ; if ( ( sumColType == VoltType . TINYINT ) || ( sumColType == VoltType . SMALLINT ) || ( sumColType == VoltType . INTEGER ) ) { return true ; } else { // no longer a good type to sum return false ; } } catch ( IllegalArgumentException e ) { // column index is bad return false ; } } | Check if the view could apply to the provided table unchanged . | 269 | 12 |
155,411 | @ Beta @ GwtIncompatible // concurrency public static ThreadFactory platformThreadFactory ( ) { if ( ! isAppEngine ( ) ) { return Executors . defaultThreadFactory ( ) ; } try { return ( ThreadFactory ) Class . forName ( "com.google_voltpatches.appengine.api.ThreadManager" ) . getMethod ( "currentRequestThreadFactory" ) . invoke ( null ) ; } catch ( IllegalAccessException e ) { throw new RuntimeException ( "Couldn't invoke ThreadManager.currentRequestThreadFactory" , e ) ; } catch ( ClassNotFoundException e ) { throw new RuntimeException ( "Couldn't invoke ThreadManager.currentRequestThreadFactory" , e ) ; } catch ( NoSuchMethodException e ) { throw new RuntimeException ( "Couldn't invoke ThreadManager.currentRequestThreadFactory" , e ) ; } catch ( InvocationTargetException e ) { throw Throwables . propagate ( e . getCause ( ) ) ; } } | Returns a default thread factory used to create new threads . | 210 | 11 |
155,412 | public static void verifySnapshots ( final List < String > directories , final Set < String > snapshotNames ) { FileFilter filter = new SnapshotFilter ( ) ; if ( ! snapshotNames . isEmpty ( ) ) { filter = new SpecificSnapshotFilter ( snapshotNames ) ; } Map < String , Snapshot > snapshots = new HashMap < String , Snapshot > ( ) ; for ( String directory : directories ) { SnapshotUtil . retrieveSnapshotFiles ( new File ( directory ) , snapshots , filter , true , SnapshotPathType . SNAP_PATH , CONSOLE_LOG ) ; } if ( snapshots . isEmpty ( ) ) { System . out . println ( "Snapshot corrupted" ) ; System . out . println ( "No files found" ) ; } for ( Snapshot s : snapshots . values ( ) ) { System . out . println ( SnapshotUtil . generateSnapshotReport ( s . getTxnId ( ) , s ) . getSecond ( ) ) ; } } | Perform snapshot verification . | 213 | 5 |
155,413 | public long lowestEquivalentValue ( final long value ) { final int bucketIndex = getBucketIndex ( value ) ; final int subBucketIndex = getSubBucketIndex ( value , bucketIndex ) ; long thisValueBaseLevel = valueFromIndex ( bucketIndex , subBucketIndex ) ; return thisValueBaseLevel ; } | Get the lowest value that is equivalent to the given value within the histogram s resolution . Where equivalent means that value samples recorded for any two equivalent values are counted in a common total count . | 70 | 38 |
155,414 | public double getMean ( ) { if ( getTotalCount ( ) == 0 ) { return 0.0 ; } recordedValuesIterator . reset ( ) ; double totalValue = 0 ; while ( recordedValuesIterator . hasNext ( ) ) { HistogramIterationValue iterationValue = recordedValuesIterator . next ( ) ; totalValue += medianEquivalentValue ( iterationValue . getValueIteratedTo ( ) ) * iterationValue . getCountAtValueIteratedTo ( ) ; } return ( totalValue * 1.0 ) / getTotalCount ( ) ; } | Get the computed mean value of all recorded values in the histogram | 118 | 13 |
155,415 | public double getStdDeviation ( ) { if ( getTotalCount ( ) == 0 ) { return 0.0 ; } final double mean = getMean ( ) ; double geometric_deviation_total = 0.0 ; recordedValuesIterator . reset ( ) ; while ( recordedValuesIterator . hasNext ( ) ) { HistogramIterationValue iterationValue = recordedValuesIterator . next ( ) ; Double deviation = ( medianEquivalentValue ( iterationValue . getValueIteratedTo ( ) ) * 1.0 ) - mean ; geometric_deviation_total += ( deviation * deviation ) * iterationValue . getCountAddedInThisIterationStep ( ) ; } double std_deviation = Math . sqrt ( geometric_deviation_total / getTotalCount ( ) ) ; return std_deviation ; } | Get the computed standard deviation of all recorded values in the histogram | 173 | 13 |
155,416 | public void reestablishTotalCount ( ) { // On overflow, the totalCount accumulated counter will (always) not match the total of counts long totalCounted = 0 ; for ( int i = 0 ; i < countsArrayLength ; i ++ ) { totalCounted += getCountAtIndex ( i ) ; } setTotalCount ( totalCounted ) ; } | Reestablish the internal notion of totalCount by recalculating it from recorded values . | 75 | 17 |
155,417 | static void setTableColumnsForSubquery ( Table table , QueryExpression queryExpression , boolean fullIndex ) { table . columnList = queryExpression . getColumns ( ) ; table . columnCount = queryExpression . getColumnCount ( ) ; table . createPrimaryKey ( ) ; if ( fullIndex ) { int [ ] colIndexes = null ; colIndexes = table . getNewColumnMap ( ) ; ArrayUtil . fillSequence ( colIndexes ) ; table . fullIndex = table . createIndexForColumns ( colIndexes ) ; } } | For table subqueries | 124 | 5 |
155,418 | public void considerCandidatePlan ( CompiledPlan plan , AbstractParsedStmt parsedStmt ) { //System.out.println(String.format("[Raw plan]:%n%s", rawplan.rootPlanGraph.toExplainPlanString())); // run the set of microptimizations, which may return many plans (or not) ScanDeterminizer . apply ( plan , m_detMode ) ; // add in the sql to the plan plan . sql = m_sql ; // compute resource usage using the single stats collector m_stats = new PlanStatistics ( ) ; AbstractPlanNode planGraph = plan . rootPlanGraph ; // compute statistics about a plan planGraph . computeEstimatesRecursively ( m_stats , m_estimates , m_paramHints ) ; // compute the cost based on the resources using the current cost model plan . cost = m_costModel . getPlanCost ( m_stats ) ; // filename for debug output String filename = String . valueOf ( m_planId ++ ) ; //* enable for debug */ System.out.println("DEBUG [new plan]: Cost:" + plan.cost + plan.rootPlanGraph.toExplainPlanString()); // find the minimum cost plan if ( m_bestPlan == null || plan . cost < m_bestPlan . cost ) { // free the PlanColumns held by the previous best plan m_bestPlan = plan ; m_bestFilename = filename ; //* enable for debug */ System.out.println("DEBUG [Best plan] updated ***\n"); } outputPlan ( plan , planGraph , filename ) ; } | Picks the best cost plan for a given raw plan | 345 | 11 |
155,419 | public static PartitionDRGateway getInstance ( int partitionId , ProducerDRGateway producerGateway , StartAction startAction ) { // if this is a primary cluster in a DR-enabled scenario // try to load the real version of this class PartitionDRGateway pdrg = null ; if ( producerGateway != null ) { pdrg = tryToLoadProVersion ( ) ; } if ( pdrg == null ) { pdrg = new PartitionDRGateway ( ) ; } // init the instance and return try { pdrg . init ( partitionId , producerGateway , startAction ) ; } catch ( Exception e ) { VoltDB . crashLocalVoltDB ( e . getMessage ( ) , true , e ) ; } // Regarding apparent lack of thread safety: this is called serially // while looping over the SPIs during database initialization assert ! m_partitionDRGateways . containsKey ( partitionId ) ; ImmutableMap . Builder < Integer , PartitionDRGateway > builder = ImmutableMap . builder ( ) ; builder . putAll ( m_partitionDRGateways ) ; builder . put ( partitionId , pdrg ) ; m_partitionDRGateways = builder . build ( ) ; return pdrg ; } | Load the full subclass if it should otherwise load the noop stub . | 272 | 14 |
155,420 | public String [ ] getUserPermissionList ( String userName ) { if ( ! m_enabled ) { return m_perm_list ; } if ( userName == null ) { return new String [ ] { } ; } AuthUser user = getUser ( userName ) ; if ( user == null ) { return new String [ ] { } ; } return user . m_permissions_list ; } | Get users permission list not god for permission checking . | 86 | 10 |
155,421 | public void callProcedure ( AuthUser user , boolean isAdmin , int timeout , ProcedureCallback cb , String procName , Object [ ] args ) { // since we know the caller, this is safe assert ( cb != null ) ; StoredProcedureInvocation task = new StoredProcedureInvocation ( ) ; task . setProcName ( procName ) ; task . setParams ( args ) ; if ( timeout != BatchTimeoutOverrideType . NO_TIMEOUT ) { task . setBatchTimeout ( timeout ) ; } InternalAdapterTaskAttributes kattrs = new InternalAdapterTaskAttributes ( DEFAULT_INTERNAL_ADAPTER_NAME , isAdmin , connectionId ( ) ) ; assert ( m_dispatcher != null ) ; // JHH: I have no idea why we need to do this, but CL crashes if we don't. Sigh. try { task = MiscUtils . roundTripForCL ( task ) ; } catch ( Exception e ) { String msg = String . format ( "Cannot invoke procedure %s. failed to create task: %s" , procName , e . getMessage ( ) ) ; m_logger . rateLimitedLog ( SUPPRESS_INTERVAL , Level . ERROR , null , msg ) ; ClientResponseImpl cri = new ClientResponseImpl ( ClientResponse . UNEXPECTED_FAILURE , new VoltTable [ 0 ] , msg ) ; try { cb . clientCallback ( cri ) ; } catch ( Exception e1 ) { throw new IllegalStateException ( e1 ) ; } } createTransaction ( kattrs , cb , task , user ) ; } | Used to call a procedure from NTPRocedureRunner Calls createTransaction with the proper params | 354 | 19 |
155,422 | public static void printSystemOut ( String message1 , long message2 ) { if ( TRACESYSTEMOUT ) { System . out . print ( message1 ) ; System . out . println ( message2 ) ; } } | Used to print messages to System . out | 48 | 8 |
155,423 | public static boolean acceptsPrecision ( int type ) { switch ( type ) { case Types . SQL_BINARY : case Types . SQL_BIT : case Types . SQL_BIT_VARYING : case Types . SQL_BLOB : case Types . SQL_CHAR : case Types . SQL_NCHAR : case Types . SQL_CLOB : case Types . NCLOB : case Types . SQL_VARBINARY : case Types . SQL_VARCHAR : case Types . SQL_NVARCHAR : case Types . VARCHAR_IGNORECASE : case Types . SQL_DECIMAL : case Types . SQL_NUMERIC : case Types . SQL_FLOAT : case Types . SQL_TIME : case Types . SQL_TIMESTAMP : case Types . SQL_INTERVAL_YEAR : case Types . SQL_INTERVAL_YEAR_TO_MONTH : case Types . SQL_INTERVAL_MONTH : case Types . SQL_INTERVAL_DAY : case Types . SQL_INTERVAL_DAY_TO_HOUR : case Types . SQL_INTERVAL_DAY_TO_MINUTE : case Types . SQL_INTERVAL_DAY_TO_SECOND : case Types . SQL_INTERVAL_HOUR : case Types . SQL_INTERVAL_HOUR_TO_MINUTE : case Types . SQL_INTERVAL_HOUR_TO_SECOND : case Types . SQL_INTERVAL_MINUTE : case Types . SQL_INTERVAL_MINUTE_TO_SECOND : case Types . SQL_INTERVAL_SECOND : case Types . VOLT_GEOGRAPHY : return true ; default : return false ; } } | Types that accept precision params in column definition or casts . CHAR VARCHAR and VARCHAR_IGNORECASE params are ignored when the sql . enforce_strict_types is false . | 361 | 40 |
155,424 | public static < T > Iterable < T > cycle ( T ... elements ) { return cycle ( Lists . newArrayList ( elements ) ) ; } | Returns an iterable whose iterators cycle indefinitely over the provided elements . | 31 | 14 |
155,425 | @ Override protected void coreLoadCatalog ( long timestamp , final byte [ ] catalogBytes ) throws EEException { LOG . trace ( "Loading Application Catalog..." ) ; int errorCode = 0 ; errorCode = nativeLoadCatalog ( pointer , timestamp , catalogBytes ) ; checkErrorCode ( errorCode ) ; //LOG.info("Loaded Catalog."); } | Provide a serialized catalog and initialize version 0 of the engine s catalog . | 74 | 16 |
155,426 | @ Override public void coreUpdateCatalog ( long timestamp , boolean isStreamUpdate , final String catalogDiffs ) throws EEException { LOG . trace ( "Loading Application Catalog..." ) ; int errorCode = 0 ; errorCode = nativeUpdateCatalog ( pointer , timestamp , isStreamUpdate , getStringBytes ( catalogDiffs ) ) ; checkErrorCode ( errorCode ) ; } | Provide a catalog diff and a new catalog version and update the engine s catalog . | 78 | 17 |
155,427 | @ Override public int extractPerFragmentStats ( int batchSize , long [ ] executionTimesOut ) { m_perFragmentStatsBuffer . clear ( ) ; // Discard the first byte since it is the timing on/off switch. m_perFragmentStatsBuffer . get ( ) ; int succeededFragmentsCount = m_perFragmentStatsBuffer . getInt ( ) ; if ( executionTimesOut != null ) { assert ( executionTimesOut . length >= succeededFragmentsCount ) ; for ( int i = 0 ; i < succeededFragmentsCount ; i ++ ) { executionTimesOut [ i ] = m_perFragmentStatsBuffer . getLong ( ) ; } // This is the time for the failed fragment. if ( succeededFragmentsCount < executionTimesOut . length ) { executionTimesOut [ succeededFragmentsCount ] = m_perFragmentStatsBuffer . getLong ( ) ; } } return succeededFragmentsCount ; } | Extract the per - fragment stats from the buffer . | 200 | 11 |
155,428 | @ Override public VoltTable [ ] getStats ( final StatsSelector selector , final int locators [ ] , final boolean interval , final Long now ) { //Clear is destructive, do it before the native call m_nextDeserializer . clear ( ) ; final int numResults = nativeGetStats ( pointer , selector . ordinal ( ) , locators , interval , now ) ; if ( numResults == - 1 ) { throwExceptionForError ( ERRORCODE_ERROR ) ; } try { m_nextDeserializer . readInt ( ) ; //Ignore the length of the result tables final VoltTable results [ ] = new VoltTable [ numResults ] ; for ( int ii = 0 ; ii < numResults ; ii ++ ) { int len = m_nextDeserializer . readInt ( ) ; byte [ ] bufCopy = new byte [ len ] ; m_nextDeserializer . readFully ( bufCopy , 0 , len ) ; // This Table should be readonly (true), but table stats need to be updated // Stream stats until Stream stats are deprecated from Table stats results [ ii ] = PrivateVoltTableFactory . createVoltTableFromBuffer ( ByteBuffer . wrap ( bufCopy ) , false ) ; } return results ; } catch ( final IOException ex ) { LOG . error ( "Failed to deserialze result table for getStats" + ex ) ; throw new EEException ( ERRORCODE_WRONG_SERIALIZED_BYTES ) ; } } | Retrieve a set of statistics using the specified selector from the StatisticsSelector enum . | 319 | 17 |
155,429 | public boolean storeLargeTempTableBlock ( long siteId , long blockCounter , ByteBuffer block ) { LargeBlockTask task = LargeBlockTask . getStoreTask ( new BlockId ( siteId , blockCounter ) , block ) ; return executeLargeBlockTaskSynchronously ( task ) ; } | Store a large temp table block to disk . | 61 | 9 |
155,430 | public boolean loadLargeTempTableBlock ( long siteId , long blockCounter , ByteBuffer block ) { LargeBlockTask task = LargeBlockTask . getLoadTask ( new BlockId ( siteId , blockCounter ) , block ) ; return executeLargeBlockTaskSynchronously ( task ) ; } | Read a large table block from disk and write it to a ByteBuffer . Block will still be stored on disk when this operation completes . | 61 | 27 |
155,431 | public boolean releaseLargeTempTableBlock ( long siteId , long blockCounter ) { LargeBlockTask task = LargeBlockTask . getReleaseTask ( new BlockId ( siteId , blockCounter ) ) ; return executeLargeBlockTaskSynchronously ( task ) ; } | Delete the block with the given id from disk . | 55 | 10 |
155,432 | public List < String > getSQLStatements ( ) { List < String > sqlStatements = new ArrayList <> ( plannedStatements . size ( ) ) ; for ( AdHocPlannedStatement plannedStatement : plannedStatements ) { sqlStatements . add ( new String ( plannedStatement . sql , Constants . UTF8ENCODING ) ) ; } return sqlStatements ; } | Retrieve all the SQL statement text as a list of strings . | 83 | 13 |
155,433 | public boolean isSinglePartitionCompatible ( ) { for ( AdHocPlannedStatement plannedStmt : plannedStatements ) { if ( plannedStmt . core . collectorFragment != null ) { return false ; } } return true ; } | Detect if batch is compatible with single partition optimizations | 52 | 9 |
155,434 | public ByteBuffer flattenPlanArrayToBuffer ( ) throws IOException { int size = 0 ; // sizeof batch ParameterSet userParamCache = null ; if ( userParamSet == null ) { userParamCache = ParameterSet . emptyParameterSet ( ) ; } else { Object [ ] typedUserParams = new Object [ userParamSet . length ] ; int ii = 0 ; for ( AdHocPlannedStatement cs : plannedStatements ) { for ( VoltType paramType : cs . core . parameterTypes ) { if ( ii >= typedUserParams . length ) { String errorMsg = "Too few actual arguments were passed for the parameters in the sql statement(s): (" + typedUserParams . length + " vs. " + ii + ")" ; // Volt-TYPE-Exception is slightly cheating, here, should there be a more general VoltArgumentException? throw new VoltTypeException ( errorMsg ) ; } typedUserParams [ ii ] = ParameterConverter . tryToMakeCompatible ( paramType . classFromType ( ) , userParamSet [ ii ] ) ; // System.out.println("DEBUG typed parameter: " + work.userParamSet[ii] + // "using type: " + paramType + "as: " + typedUserParams[ii]); ii ++ ; } } // Each parameter referenced in each statements should be represented // exactly once in userParams. if ( ii < typedUserParams . length ) { // Volt-TYPE-Exception is slightly cheating, here, should there be a more general VoltArgumentException? String errorMsg = "Too many actual arguments were passed for the parameters in the sql statement(s): (" + typedUserParams . length + " vs. " + ii + ")" ; throw new VoltTypeException ( errorMsg ) ; } userParamCache = ParameterSet . fromArrayNoCopy ( typedUserParams ) ; } size += userParamCache . getSerializedSize ( ) ; size += 2 ; // sizeof batch for ( AdHocPlannedStatement cs : plannedStatements ) { size += cs . getSerializedSize ( ) ; } ByteBuffer buf = ByteBuffer . allocate ( size ) ; userParamCache . flattenToBuffer ( buf ) ; buf . putShort ( ( short ) plannedStatements . size ( ) ) ; for ( AdHocPlannedStatement cs : plannedStatements ) { cs . flattenToBuffer ( buf ) ; } return buf ; } | For convenience serialization is accomplished with this single method but deserialization is piecemeal via the static methods userParamsFromBuffer and planArrayFromBuffer with no dummy AdHocPlannedStmtBatch receiver instance required . | 525 | 47 |
155,435 | public String explainStatement ( int i , Database db , boolean getJSONString ) { AdHocPlannedStatement plannedStatement = plannedStatements . get ( i ) ; String aggplan = new String ( plannedStatement . core . aggregatorFragment , Constants . UTF8ENCODING ) ; PlanNodeTree pnt = new PlanNodeTree ( ) ; try { String result = null ; JSONObject jobj = new JSONObject ( aggplan ) ; if ( getJSONString ) { result = jobj . toString ( 4 ) ; } pnt . loadFromJSONPlan ( jobj , db ) ; if ( plannedStatement . core . collectorFragment != null ) { // multi-partition query plan String collplan = new String ( plannedStatement . core . collectorFragment , Constants . UTF8ENCODING ) ; PlanNodeTree collpnt = new PlanNodeTree ( ) ; // reattach plan fragments JSONObject jobMP = new JSONObject ( collplan ) ; collpnt . loadFromJSONPlan ( jobMP , db ) ; assert ( collpnt . getRootPlanNode ( ) instanceof SendPlanNode ) ; pnt . getRootPlanNode ( ) . reattachFragment ( collpnt . getRootPlanNode ( ) ) ; if ( getJSONString ) { result += "\n" + jobMP . toString ( 4 ) ; } } if ( ! getJSONString ) { result = pnt . getRootPlanNode ( ) . toExplainPlanString ( ) ; } return result ; } catch ( JSONException e ) { System . out . println ( e ) ; return "Internal Error (JSONException): " + e . getMessage ( ) ; } } | Return the EXPLAIN string of the batched statement at the index | 360 | 14 |
155,436 | public synchronized static AdHocCompilerCache getCacheForCatalogHash ( byte [ ] catalogHash ) { String hashString = Encoder . hexEncode ( catalogHash ) ; AdHocCompilerCache cache = m_catalogHashMatch . getIfPresent ( hashString ) ; if ( cache == null ) { cache = new AdHocCompilerCache ( ) ; m_catalogHashMatch . put ( hashString , cache ) ; } return cache ; } | Get the global cache for a given hash of the catalog . Note that there can be only one cache per catalogHash at a time . | 99 | 27 |
155,437 | synchronized void printStats ( ) { String line1 = String . format ( "CACHE STATS - Literals: Hits %d/%d (%.1f%%), Inserts %d Evictions %d\n" , m_literalHits , m_literalQueries , ( m_literalHits * 100.0 ) / m_literalQueries , m_literalInsertions , m_literalEvictions ) ; String line2 = String . format ( "CACHE STATS - Plans: Hits %d/%d (%.1f%%), Inserts %d Evictions %d\n" , m_planHits , m_planQueries , ( m_planHits * 100.0 ) / m_planQueries , m_planInsertions , m_planEvictions ) ; System . out . print ( line1 + line2 ) ; System . out . flush ( ) ; // reset these m_literalHits = 0 ; m_literalQueries = 0 ; m_literalInsertions = 0 ; m_literalEvictions = 0 ; m_planHits = 0 ; m_planQueries = 0 ; m_planInsertions = 0 ; m_planEvictions = 0 ; } | Stats printing method used during development . Probably shouldn t live past real stats integration . | 275 | 16 |
155,438 | public synchronized void put ( String sql , String parsedToken , AdHocPlannedStatement planIn , String [ ] extractedLiterals , boolean hasUserQuestionMarkParameters , boolean hasAutoParameterizedException ) { assert ( sql != null ) ; assert ( parsedToken != null ) ; assert ( planIn != null ) ; AdHocPlannedStatement plan = planIn ; assert ( new String ( plan . sql , Constants . UTF8ENCODING ) . equals ( sql ) ) ; // hasUserQuestionMarkParameters and hasAutoParameterizedException can not be true at the same time // it means that a query can not be both user parameterized query and auto parameterized query. assert ( ! hasUserQuestionMarkParameters || ! hasAutoParameterizedException ) ; // uncomment this to get some raw stdout cache performance stats every 5s //startPeriodicStatsPrinting(); // deal with L2 cache if ( ! hasAutoParameterizedException ) { BoundPlan matched = null ; BoundPlan unmatched = new BoundPlan ( planIn . core , planIn . parameterBindings ( extractedLiterals ) ) ; // deal with the parameterized plan cache first List < BoundPlan > boundVariants = m_coreCache . get ( parsedToken ) ; if ( boundVariants == null ) { boundVariants = new ArrayList < BoundPlan > ( ) ; m_coreCache . put ( parsedToken , boundVariants ) ; // Note that there is an edge case in which more than one plan is getting counted as one // "plan insertion". This only happens when two different plans arose from the same parameterized // query (token) because one invocation used the correct constants to trigger an expression index and // another invocation did not. These are not counted separately (which would have to happen below // after each call to boundVariants.add) because they are not evicted separately. // It seems saner to use consistent units when counting insertions vs. evictions. ++ m_planInsertions ; } else { for ( BoundPlan boundPlan : boundVariants ) { if ( boundPlan . equals ( unmatched ) ) { matched = boundPlan ; break ; } } if ( matched != null ) { // if a different core is found, reuse it // this is useful when updating the literal cache if ( unmatched . m_core != matched . m_core ) { plan = new AdHocPlannedStatement ( planIn , matched . m_core ) ; plan . setBoundConstants ( matched . m_constants ) ; } } } if ( matched == null ) { // Don't count insertions (of possibly repeated tokens) here // -- see the comment above where only UNIQUE token insertions are being counted, instead. boundVariants . add ( unmatched ) ; } } // then deal with the L1 cache if ( ! hasUserQuestionMarkParameters ) { AdHocPlannedStatement cachedPlan = m_literalCache . get ( sql ) ; if ( cachedPlan == null ) { //* enable to debug */ System.out.println("DEBUG: Caching literal '" + sql + "'"); m_literalCache . put ( sql , plan ) ; ++ m_literalInsertions ; } else { assert ( cachedPlan . equals ( plan ) ) ; } } } | Called from the PlannerTool directly when it finishes planning . This is the only way to populate the cache . | 689 | 23 |
155,439 | public void startPeriodicStatsPrinting ( ) { if ( m_statsTimer == null ) { m_statsTimer = new Timer ( ) ; m_statsTimer . scheduleAtFixedRate ( new TimerTask ( ) { @ Override public void run ( ) { printStats ( ) ; } } , 5000 , 5000 ) ; } } | Start a timer that prints cache stats to the console every 5s . Used for development until we get better stats integration . | 74 | 24 |
155,440 | @ Override public Date getDate ( int parameterIndex , Calendar cal ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Retrieves the value of the designated JDBC DATE parameter as a java . sql . Date object using the given Calendar object to construct the date . | 39 | 31 |
155,441 | @ Override public Date getDate ( String parameterName , Calendar cal ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Retrieves the value of a JDBC DATE parameter as a java . sql . Date object using the given Calendar object to construct the date . | 39 | 30 |
155,442 | @ Override public Object getObject ( String parameterName , Map < String , Class < ? > > map ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Returns an object representing the value of OUT parameter parameterName and uses map for the custom mapping of the parameter value . | 47 | 23 |
155,443 | @ Override public Time getTime ( int parameterIndex , Calendar cal ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Retrieves the value of the designated JDBC TIME parameter as a java . sql . Time object using the given Calendar object to construct the time . | 39 | 30 |
155,444 | @ Override public Time getTime ( String parameterName , Calendar cal ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Retrieves the value of a JDBC TIME parameter as a java . sql . Time object using the given Calendar object to construct the time . | 39 | 29 |
155,445 | @ Override public Timestamp getTimestamp ( int parameterIndex , Calendar cal ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Retrieves the value of the designated JDBC TIMESTAMP parameter as a java . sql . Timestamp object using the given Calendar object to construct the Timestamp object . | 41 | 35 |
155,446 | @ Override public Timestamp getTimestamp ( String parameterName , Calendar cal ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Retrieves the value of a JDBC TIMESTAMP parameter as a java . sql . Timestamp object using the given Calendar object to construct the Timestamp object . | 41 | 34 |
155,447 | @ Override public void registerOutParameter ( String parameterName , int sqlType , int scale ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Registers the parameter named parameterName to be of JDBC type sqlType . | 44 | 16 |
155,448 | @ Override public void setNString ( String parameterName , String value ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Sets the designated parameter to the given String object . | 40 | 11 |
155,449 | @ Override public void setURL ( String parameterName , URL val ) throws SQLException { checkClosed ( ) ; throw SQLError . noSupport ( ) ; } | Sets the designated parameter to the given java . net . URL object . | 39 | 15 |
155,450 | public static String suffixHSIdsWithMigratePartitionLeaderRequest ( Long HSId ) { return Long . toString ( Long . MAX_VALUE ) + "/" + Long . toString ( HSId ) + migrate_partition_leader_suffix ; } | Generate a HSID string with BALANCE_SPI_SUFFIX information . When this string is updated we can tell the reason why HSID is changed . | 56 | 34 |
155,451 | public void startPartitionWatch ( ) throws InterruptedException , ExecutionException { Future < ? > task = m_es . submit ( new PartitionWatchEvent ( ) ) ; task . get ( ) ; } | Initialized and start watching partition level cache this function is blocking . | 44 | 12 |
155,452 | private void processPartitionWatchEvent ( ) throws KeeperException , InterruptedException { try { m_zk . create ( m_rootNode , null , Ids . OPEN_ACL_UNSAFE , CreateMode . PERSISTENT ) ; m_zk . getData ( m_rootNode , m_childWatch , null ) ; } catch ( KeeperException . NodeExistsException e ) { m_zk . getData ( m_rootNode , m_childWatch , null ) ; } } | Race to create partition - specific zk node and put a watch on it . | 111 | 16 |
155,453 | public Object convertToDefaultType ( SessionInterface session , Object a ) { if ( a == null ) { return a ; } Type otherType ; if ( a instanceof Number ) { if ( a instanceof BigInteger ) { a = new BigDecimal ( ( BigInteger ) a ) ; } else if ( a instanceof Float ) { a = new Double ( ( ( Float ) a ) . doubleValue ( ) ) ; } else if ( a instanceof Byte ) { a = ValuePool . getInt ( ( ( Byte ) a ) . intValue ( ) ) ; } else if ( a instanceof Short ) { a = ValuePool . getInt ( ( ( Short ) a ) . intValue ( ) ) ; } if ( a instanceof Integer ) { otherType = Type . SQL_INTEGER ; } else if ( a instanceof Long ) { otherType = Type . SQL_BIGINT ; } else if ( a instanceof Double ) { otherType = Type . SQL_DOUBLE ; } else if ( a instanceof BigDecimal ) { // BEGIN Cherry-picked code change from hsqldb-2.2.8 otherType = Type . SQL_DECIMAL_DEFAULT ; /* if (typeCode == Types.SQL_DECIMAL || typeCode == Types.SQL_NUMERIC) { return convertToTypeLimits(session, a); } BigDecimal val = (BigDecimal) a; otherType = getNumberType(Types.SQL_DECIMAL, JavaSystem.precision(val), scale); */ // END Cherry-picked code change from hsqldb-2.2.8 } else { throw Error . error ( ErrorCode . X_42561 ) ; } // BEGIN Cherry-picked code change from hsqldb-2.2.8 switch ( typeCode ) { case Types . TINYINT : case Types . SQL_SMALLINT : case Types . SQL_INTEGER : return convertToInt ( session , a , Types . INTEGER ) ; case Types . SQL_BIGINT : return convertToLong ( session , a ) ; case Types . SQL_REAL : case Types . SQL_FLOAT : case Types . SQL_DOUBLE : return convertToDouble ( a ) ; case Types . SQL_NUMERIC : case Types . SQL_DECIMAL : { a = convertToDecimal ( a ) ; BigDecimal dec = ( BigDecimal ) a ; if ( scale != dec . scale ( ) ) { dec = dec . setScale ( scale , BigDecimal . ROUND_HALF_DOWN ) ; } return dec ; } default : throw Error . error ( ErrorCode . X_42561 ) ; } // END Cherry-picked code change from hsqldb-2.2.8 } else if ( a instanceof String ) { otherType = Type . SQL_VARCHAR ; } else { throw Error . error ( ErrorCode . X_42561 ) ; } return convertToType ( session , a , otherType ) ; } | Converts a value to this type | 658 | 7 |
155,454 | private static Double convertToDouble ( Object a ) { double value ; if ( a instanceof java . lang . Double ) { return ( Double ) a ; } else if ( a instanceof BigDecimal ) { BigDecimal bd = ( BigDecimal ) a ; value = bd . doubleValue ( ) ; int signum = bd . signum ( ) ; BigDecimal bdd = new BigDecimal ( value + signum ) ; if ( bdd . compareTo ( bd ) != signum ) { throw Error . error ( ErrorCode . X_22003 ) ; } } else { value = ( ( Number ) a ) . doubleValue ( ) ; } return ValuePool . getDouble ( Double . doubleToLongBits ( value ) ) ; } | Converter from a numeric object to Double . Input is checked to be within range represented by Double | 165 | 20 |
155,455 | public Object mod ( Object a , Object b ) { if ( a == null || b == null ) { return null ; } switch ( typeCode ) { case Types . SQL_REAL : case Types . SQL_FLOAT : case Types . SQL_DOUBLE : { double ad = ( ( Number ) a ) . doubleValue ( ) ; double bd = ( ( Number ) b ) . doubleValue ( ) ; if ( bd == 0 ) { throw Error . error ( ErrorCode . X_22012 ) ; } return ValuePool . getDouble ( Double . doubleToLongBits ( ad % bd ) ) ; } case Types . SQL_DECIMAL : { if ( ( b ) . equals ( 0 ) ) { throw Error . error ( ErrorCode . X_22012 ) ; } return ValuePool . getBigDecimal ( ( ( BigDecimal ) a ) . remainder ( ( BigDecimal ) b ) ) ; } case Types . TINYINT : case Types . SQL_SMALLINT : case Types . SQL_INTEGER : { int ai = ( ( Number ) a ) . intValue ( ) ; int bi = ( ( Number ) b ) . intValue ( ) ; if ( bi == 0 ) { throw Error . error ( ErrorCode . X_22012 ) ; } return ValuePool . getInt ( ai % bi ) ; } case Types . SQL_BIGINT : { long al = ( ( Number ) a ) . longValue ( ) ; long bl = ( ( Number ) b ) . longValue ( ) ; if ( bl == 0 ) { throw Error . error ( ErrorCode . X_22012 ) ; } return ValuePool . getLong ( al % bl ) ; } default : throw Error . runtimeError ( ErrorCode . U_S0500 , "NumberType" ) ; } } | A VoltDB extension | 397 | 4 |
155,456 | public synchronized void write ( int c ) throws IOException { checkClosed ( ) ; int newcount = count + 1 ; if ( newcount > buf . length ) { buf = copyOf ( buf , Math . max ( buf . length << 1 , newcount ) ) ; } buf [ count ] = ( char ) c ; count = newcount ; } | Writes the specified single character . | 75 | 7 |
155,457 | private void initiateSPIMigrationIfRequested ( Iv2InitiateTaskMessage msg ) { if ( ! "@MigratePartitionLeader" . equals ( msg . getStoredProcedureName ( ) ) ) { return ; } final Object [ ] params = msg . getParameters ( ) ; int pid = Integer . parseInt ( params [ 1 ] . toString ( ) ) ; if ( pid != m_partitionId ) { tmLog . warn ( String . format ( "@MigratePartitionLeader executed at a wrong partition %d for partition %d." , m_partitionId , pid ) ) ; return ; } RealVoltDB db = ( RealVoltDB ) VoltDB . instance ( ) ; int hostId = Integer . parseInt ( params [ 2 ] . toString ( ) ) ; Long newLeaderHSId = db . getCartographer ( ) . getHSIDForPartitionHost ( hostId , pid ) ; if ( newLeaderHSId == null || newLeaderHSId == m_hsId ) { tmLog . warn ( String . format ( "@MigratePartitionLeader the partition leader is already on the host %d or the host id is invalid." , hostId ) ) ; return ; } SpScheduler scheduler = ( SpScheduler ) m_scheduler ; scheduler . checkPointMigratePartitionLeader ( ) ; scheduler . m_isLeader = false ; m_newLeaderHSID = newLeaderHSId ; m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . STARTED ; LeaderCache leaderAppointee = new LeaderCache ( m_messenger . getZK ( ) , "initiateSPIMigrationIfRequested-" + m_partitionId , VoltZK . iv2appointees ) ; try { leaderAppointee . start ( true ) ; leaderAppointee . put ( pid , LeaderCache . suffixHSIdsWithMigratePartitionLeaderRequest ( newLeaderHSId ) ) ; } catch ( InterruptedException | ExecutionException | KeeperException e ) { VoltDB . crashLocalVoltDB ( "fail to start MigratePartitionLeader" , true , e ) ; } finally { try { leaderAppointee . shutdown ( ) ; } catch ( InterruptedException e ) { } } tmLog . info ( "MigratePartitionLeader for partition " + pid + " to " + CoreUtils . hsIdToString ( newLeaderHSId ) ) ; //notify the new leader right away if the current leader has drained all transactions. notifyNewLeaderOfTxnDoneIfNeeded ( ) ; } | rerouted from this moment on until the transactions are correctly routed to new leader . | 566 | 16 |
155,458 | private boolean checkMisroutedIv2IntiateTaskMessage ( Iv2InitiateTaskMessage message ) { if ( message . isForReplica ( ) ) { return false ; } if ( m_scheduler . isLeader ( ) && m_migratePartitionLeaderStatus != MigratePartitionLeaderStatus . TXN_RESTART ) { //At this point, the message is sent to partition leader return false ; } //At this point, the message is misrouted. //(1) If a site has been demoted via @MigratePartitionLeader, the messages which are sent to the leader will be restarted. //(2) If a site becomes new leader via @MigratePartitionLeader. Transactions will be restarted before it gets notification from old // leader that transactions on older leader have been drained. InitiateResponseMessage response = new InitiateResponseMessage ( message ) ; response . setMisrouted ( message . getStoredProcedureInvocation ( ) ) ; response . m_sourceHSId = getHSId ( ) ; deliver ( response ) ; if ( tmLog . isDebugEnabled ( ) ) { tmLog . debug ( "Sending message back on:" + CoreUtils . hsIdToString ( m_hsId ) + " isLeader:" + m_scheduler . isLeader ( ) + " status:" + m_migratePartitionLeaderStatus + "\n" + message ) ; } //notify the new partition leader that the old leader has completed the Txns if needed. notifyNewLeaderOfTxnDoneIfNeeded ( ) ; return true ; } | if these requests are intended for leader . Client interface will restart these transactions . | 348 | 15 |
155,459 | private boolean checkMisroutedFragmentTaskMessage ( FragmentTaskMessage message ) { if ( m_scheduler . isLeader ( ) || message . isForReplica ( ) ) { return false ; } TransactionState txnState = ( ( ( SpScheduler ) m_scheduler ) . getTransactionState ( message . getTxnId ( ) ) ) ; // If a fragment is part of a transaction which have not been seen on this site, restart it. if ( txnState == null ) { FragmentResponseMessage response = new FragmentResponseMessage ( message , getHSId ( ) ) ; TransactionRestartException restart = new TransactionRestartException ( "Transaction being restarted due to MigratePartitionLeader." , message . getTxnId ( ) ) ; restart . setMisrouted ( true ) ; response . setStatus ( FragmentResponseMessage . UNEXPECTED_ERROR , restart ) ; response . m_sourceHSId = getHSId ( ) ; response . setPartitionId ( m_partitionId ) ; if ( tmLog . isDebugEnabled ( ) ) { tmLog . debug ( "misRoutedFragMsg on site:" + CoreUtils . hsIdToString ( getHSId ( ) ) + "\n" + message ) ; } deliver ( response ) ; return true ; } // A transaction may have multiple batches or fragments. If the first batch or fragment has already been // processed, the follow-up batches or fragments should also be processed on this site. if ( ! m_scheduler . isLeader ( ) && ! message . isForReplica ( ) ) { message . setExecutedOnPreviousLeader ( true ) ; txnState . setLeaderMigrationInvolved ( ) ; if ( tmLog . isDebugEnabled ( ) ) { tmLog . debug ( "Follow-up fragment will be processed on " + CoreUtils . hsIdToString ( getHSId ( ) ) + "\n" + message ) ; } } return false ; } | After MigratePartitionLeader has been requested the fragments which are sent to leader site should be restarted . | 436 | 22 |
155,460 | private void handleLogRequest ( VoltMessage message ) { Iv2RepairLogRequestMessage req = ( Iv2RepairLogRequestMessage ) message ; // It is possible for a dead host to queue messages after a repair request is processed // so make sure this can't happen by re-queuing this message after we know the dead host is gone // Since we are not checking validateForeignHostId on the PicoNetwork thread, it is possible for // the PicoNetwork thread to validateForeignHostId and queue a message behind this repair message. // Further, we loose visibility to the ForeignHost as soon as HostMessenger marks the host invalid // even though the PicoNetwork thread could still be alive so we will skeptically int deadHostId = req . getDeadHostId ( ) ; if ( deadHostId != Integer . MAX_VALUE ) { if ( m_messenger . canCompleteRepair ( deadHostId ) ) { // Make sure we are the last in the task queue when we know the ForeignHost is gone req . disableDeadHostCheck ( ) ; deliver ( message ) ; } else { if ( req . getRepairRetryCount ( ) > 100 && req . getRepairRetryCount ( ) % 100 == 0 ) { hostLog . warn ( "Repair Request for dead host " + deadHostId + " has not been processed yet because connection has not closed" ) ; } Runnable retryRepair = new Runnable ( ) { @ Override public void run ( ) { InitiatorMailbox . this . deliver ( message ) ; } } ; VoltDB . instance ( ) . scheduleWork ( retryRepair , 10 , - 1 , TimeUnit . MILLISECONDS ) ; // the repair message will be resubmitted shortly when the ForeignHosts to the dead host have been removed } return ; } List < Iv2RepairLogResponseMessage > logs = m_repairLog . contents ( req . getRequestId ( ) , req . isMPIRequest ( ) ) ; if ( req . isMPIRequest ( ) ) { m_scheduler . cleanupTransactionBacklogOnRepair ( ) ; } for ( Iv2RepairLogResponseMessage log : logs ) { send ( message . m_sourceHSId , log ) ; } } | Produce the repair log . This is idempotent . | 484 | 13 |
155,461 | private void setMigratePartitionLeaderStatus ( MigratePartitionLeaderMessage message ) { //The host with old partition leader is down. if ( message . isStatusReset ( ) ) { m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . NONE ; return ; } if ( m_migratePartitionLeaderStatus == MigratePartitionLeaderStatus . NONE ) { //txn draining notification from the old leader arrives before this site is promoted m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . TXN_DRAINED ; } else if ( m_migratePartitionLeaderStatus == MigratePartitionLeaderStatus . TXN_RESTART ) { //if the new leader has been promoted, stop restarting txns. m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . NONE ; } tmLog . info ( "MigratePartitionLeader new leader " + CoreUtils . hsIdToString ( m_hsId ) + " is notified by previous leader " + CoreUtils . hsIdToString ( message . getPriorLeaderHSID ( ) ) + ". status:" + m_migratePartitionLeaderStatus ) ; } | that previous partition leader has drained its txns | 263 | 9 |
155,462 | public void setMigratePartitionLeaderStatus ( boolean migratePartitionLeader ) { if ( ! migratePartitionLeader ) { m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . NONE ; m_newLeaderHSID = Long . MIN_VALUE ; return ; } //The previous leader has already drained all txns if ( m_migratePartitionLeaderStatus == MigratePartitionLeaderStatus . TXN_DRAINED ) { m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . NONE ; tmLog . info ( "MigratePartitionLeader transactions on previous partition leader are drained. New leader:" + CoreUtils . hsIdToString ( m_hsId ) + " status:" + m_migratePartitionLeaderStatus ) ; return ; } //Wait for the notification from old partition leader m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . TXN_RESTART ; tmLog . info ( "MigratePartitionLeader restart txns on new leader:" + CoreUtils . hsIdToString ( m_hsId ) + " status:" + m_migratePartitionLeaderStatus ) ; } | the site for new partition leader | 257 | 6 |
155,463 | public void notifyNewLeaderOfTxnDoneIfNeeded ( ) { //return quickly to avoid performance hit if ( m_newLeaderHSID == Long . MIN_VALUE ) { return ; } SpScheduler scheduler = ( SpScheduler ) m_scheduler ; if ( ! scheduler . txnDoneBeforeCheckPoint ( ) ) { return ; } MigratePartitionLeaderMessage message = new MigratePartitionLeaderMessage ( m_hsId , m_newLeaderHSID ) ; send ( message . getNewLeaderHSID ( ) , message ) ; //reset status on the old partition leader m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . NONE ; m_repairLog . setLeaderState ( false ) ; tmLog . info ( "MigratePartitionLeader previous leader " + CoreUtils . hsIdToString ( m_hsId ) + " notifies new leader " + CoreUtils . hsIdToString ( m_newLeaderHSID ) + " transactions are drained." + " status:" + m_migratePartitionLeaderStatus ) ; m_newLeaderHSID = Long . MIN_VALUE ; } | Then new master can proceed to process transactions . | 253 | 9 |
155,464 | public void resetMigratePartitionLeaderStatus ( ) { m_scheduler . m_isLeader = true ; m_migratePartitionLeaderStatus = MigratePartitionLeaderStatus . NONE ; m_repairLog . setLeaderState ( true ) ; m_newLeaderHSID = Long . MIN_VALUE ; } | Reinstall the site as leader . | 70 | 7 |
155,465 | private Option resolveOption ( String opt ) { opt = Util . stripLeadingHyphens ( opt ) ; for ( Option option : options ) { if ( opt . equals ( option . getOpt ( ) ) ) { return option ; } if ( opt . equals ( option . getLongOpt ( ) ) ) { return option ; } } return null ; } | Retrieves the option object given the long or short option as a String | 75 | 15 |
155,466 | public String [ ] getArgs ( ) { String [ ] answer = new String [ args . size ( ) ] ; args . toArray ( answer ) ; return answer ; } | Retrieve any left - over non - recognized options and arguments | 36 | 12 |
155,467 | public boolean processScanNodeWithReAggNode ( AbstractPlanNode node , AbstractPlanNode reAggNode ) { // MV table scan node can not be in in-lined nodes. for ( int i = 0 ; i < node . getChildCount ( ) ; i ++ ) { AbstractPlanNode child = node . getChild ( i ) ; if ( child instanceof AbstractScanPlanNode ) { AbstractScanPlanNode scanNode = ( AbstractScanPlanNode ) child ; if ( ! scanNode . getTargetTableName ( ) . equals ( getMVTableName ( ) ) ) { continue ; } if ( reAggNode != null ) { // Join query case. node . setAndLinkChild ( i , reAggNode ) ; } // Process scan node. // Set up the scan plan node's scan columns. Add in-line projection node for scan node. scanNode . addInlinePlanNode ( m_scanInlinedProjectionNode ) ; m_scanNode = scanNode ; return true ; } else { boolean replaced = processScanNodeWithReAggNode ( child , reAggNode ) ; if ( replaced ) { return true ; } } } return false ; } | Find the scan node on MV table replace it with reAggNode for join query . This scan node can not be in - lined so it should be as a child of a join node . | 246 | 38 |
155,468 | private void resolveColumnReferences ( ) { if ( isDistinctSelect || isGrouped ) { acceptsSequences = false ; } for ( int i = 0 ; i < rangeVariables . length ; i ++ ) { Expression e = rangeVariables [ i ] . nonIndexJoinCondition ; if ( e == null ) { continue ; } resolveColumnReferencesAndAllocate ( e , i + 1 , false ) ; } resolveColumnReferencesAndAllocate ( queryCondition , rangeVariables . length , false ) ; for ( int i = 0 ; i < indexLimitVisible ; i ++ ) { resolveColumnReferencesAndAllocate ( exprColumns [ i ] , rangeVariables . length , acceptsSequences ) ; } for ( int i = indexLimitVisible ; i < indexStartOrderBy ; i ++ ) { resolveColumnReferencesAndAllocate ( exprColumns [ i ] , rangeVariables . length , false ) ; } /************************* Volt DB Extensions *************************/ resolveColumnReferencesInGroupBy ( ) ; /**********************************************************************/ resolveColumnReferencesInOrderBy ( sortAndSlice ) ; } | Resolves all column expressions in the GROUP BY clause and beyond . Replaces any alias column expression in the ORDER BY cluase with the actual select column expression . | 236 | 33 |
155,469 | private int getMaxRowCount ( Session session , int rowCount ) { int limitStart = getLimitStart ( session ) ; int limitCount = getLimitCount ( session , rowCount ) ; if ( simpleLimit ) { if ( rowCount == 0 ) { rowCount = limitCount ; } // A VoltDB extension to support LIMIT 0 if ( rowCount > Integer . MAX_VALUE - limitStart ) { /* disable 1 line ... if (rowCount == 0 || rowCount > Integer.MAX_VALUE - limitStart) { ... disabled 1 line */ // End of VoltDB extension rowCount = Integer . MAX_VALUE ; } else { rowCount += limitStart ; } } else { rowCount = Integer . MAX_VALUE ; // A VoltDB extension to support LIMIT 0 // limitCount == 0 can be enforced/optimized as rowCount == 0 regardless of offset // even in non-simpleLimit cases (SELECT DISTINCT, GROUP BY, and/or ORDER BY). // This is an optimal handling of a hard-coded LIMIT 0, but it really shouldn't be the ONLY // enforcement for zero LIMITs -- what about "LIMIT ?" with 0 passed later as a parameter? // The HSQL executor ("HSQL back end") also needs runtime enforcement of zero limits. // The VoltDB executor has such enforcement. if ( limitCount == 0 ) { rowCount = 0 ; } // End of VoltDB extension } return rowCount ; } | translate the rowCount into total number of rows needed from query including any rows skipped at the beginning | 309 | 20 |
155,470 | protected void dumpExprColumns ( String header ) { System . out . println ( "\n\n*********************************************" ) ; System . out . println ( header ) ; try { System . out . println ( getSQL ( ) ) ; } catch ( Exception e ) { } for ( int i = 0 ; i < exprColumns . length ; ++ i ) { if ( i == 0 ) System . out . println ( "Visible columns:" ) ; if ( i == indexStartOrderBy ) System . out . println ( "start order by:" ) ; if ( i == indexStartAggregates ) System . out . println ( "start aggregates:" ) ; if ( i == indexLimitVisible ) System . out . println ( "After limit of visible columns:" ) ; System . out . println ( i + ": " + exprColumns [ i ] ) ; } System . out . println ( "\n\n" ) ; } | Dumps the exprColumns list for this query specification . Writes to stdout . | 201 | 18 |
155,471 | public void updateEECacheStats ( long eeCacheSize , long hits , long misses , int partitionId ) { m_cache1Level = eeCacheSize ; m_cache1Hits += hits ; m_cacheMisses += misses ; m_invocations += hits + misses ; m_partitionId = partitionId ; } | Used to update EE cache stats without changing tracked time | 72 | 10 |
155,472 | public void endStatsCollection ( long cache1Size , long cache2Size , CacheUse cacheUse , long partitionId ) { if ( m_currentStartTime != null ) { long delta = System . nanoTime ( ) - m_currentStartTime ; if ( delta < 0 ) { if ( Math . abs ( delta ) > 1000000000 ) { log . info ( "Planner statistics recorded a negative planning time larger than one second: " + delta ) ; } } else { m_totalPlanningTime += delta ; m_minPlanningTime = Math . min ( delta , m_minPlanningTime ) ; m_maxPlanningTime = Math . max ( delta , m_maxPlanningTime ) ; m_lastMinPlanningTime = Math . min ( delta , m_lastMinPlanningTime ) ; m_lastMaxPlanningTime = Math . max ( delta , m_lastMaxPlanningTime ) ; } m_currentStartTime = null ; } m_cache1Level = cache1Size ; m_cache2Level = cache2Size ; switch ( cacheUse ) { case HIT1 : m_cache1Hits ++ ; break ; case HIT2 : m_cache2Hits ++ ; break ; case MISS : m_cacheMisses ++ ; break ; case FAIL : m_failures ++ ; break ; } m_invocations ++ ; m_partitionId = partitionId ; } | Called after planning or failing to plan . Records timer and cache stats . | 302 | 15 |
155,473 | @ Override protected void updateStatsRow ( Object rowKey , Object rowValues [ ] ) { super . updateStatsRow ( rowKey , rowValues ) ; rowValues [ columnNameToIndex . get ( "PARTITION_ID" ) ] = m_partitionId ; long totalTimedExecutionTime = m_totalPlanningTime ; long minExecutionTime = m_minPlanningTime ; long maxExecutionTime = m_maxPlanningTime ; long cache1Level = m_cache1Level ; long cache2Level = m_cache2Level ; long cache1Hits = m_cache1Hits ; long cache2Hits = m_cache2Hits ; long cacheMisses = m_cacheMisses ; long failureCount = m_failures ; if ( m_interval ) { totalTimedExecutionTime = m_totalPlanningTime - m_lastTimedPlanningTime ; m_lastTimedPlanningTime = m_totalPlanningTime ; minExecutionTime = m_lastMinPlanningTime ; maxExecutionTime = m_lastMaxPlanningTime ; m_lastMinPlanningTime = Long . MAX_VALUE ; m_lastMaxPlanningTime = Long . MIN_VALUE ; cache1Level = m_cache1Level - m_lastCache1Level ; m_lastCache1Level = m_cache1Level ; cache2Level = m_cache2Level - m_lastCache2Level ; m_lastCache2Level = m_cache2Level ; cache1Hits = m_cache1Hits - m_lastCache1Hits ; m_lastCache1Hits = m_cache1Hits ; cache2Hits = m_cache2Hits - m_lastCache2Hits ; m_lastCache2Hits = m_cache2Hits ; cacheMisses = m_cacheMisses - m_lastCacheMisses ; m_lastCacheMisses = m_cacheMisses ; failureCount = m_failures - m_lastFailures ; m_lastFailures = m_failures ; m_lastInvocations = m_invocations ; } rowValues [ columnNameToIndex . get ( VoltSystemProcedure . CNAME_SITE_ID ) ] = m_siteId ; rowValues [ columnNameToIndex . get ( "PARTITION_ID" ) ] = m_partitionId ; rowValues [ columnNameToIndex . get ( "CACHE1_LEVEL" ) ] = cache1Level ; rowValues [ columnNameToIndex . get ( "CACHE2_LEVEL" ) ] = cache2Level ; rowValues [ columnNameToIndex . get ( "CACHE1_HITS" ) ] = cache1Hits ; rowValues [ columnNameToIndex . get ( "CACHE2_HITS" ) ] = cache2Hits ; rowValues [ columnNameToIndex . get ( "CACHE_MISSES" ) ] = cacheMisses ; rowValues [ columnNameToIndex . get ( "PLAN_TIME_MIN" ) ] = minExecutionTime ; rowValues [ columnNameToIndex . get ( "PLAN_TIME_MAX" ) ] = maxExecutionTime ; if ( getSampleCount ( ) != 0 ) { rowValues [ columnNameToIndex . get ( "PLAN_TIME_AVG" ) ] = ( totalTimedExecutionTime / getSampleCount ( ) ) ; } else { rowValues [ columnNameToIndex . get ( "PLAN_TIME_AVG" ) ] = 0L ; } rowValues [ columnNameToIndex . get ( "FAILURES" ) ] = failureCount ; } | Update the rowValues array with the latest statistical information . This method is overrides the super class version which must also be called so that it can update its columns . | 805 | 33 |
155,474 | static void tag ( StringBuilder sb , String color , String text ) { sb . append ( "<span class='label label" ) ; if ( color != null ) { sb . append ( "-" ) . append ( color ) ; } String classText = text . replace ( ' ' , ' ' ) ; sb . append ( " l-" ) . append ( classText ) . append ( "'>" ) . append ( text ) . append ( "</span>" ) ; } | Make an html bootstrap tag with our custom css class . | 103 | 13 |
155,475 | public static String report ( Catalog catalog , long minHeap , boolean isPro , int hostCount , int sitesPerHost , int kfactor , ArrayList < Feedback > warnings , String autoGenDDL ) throws IOException { // asynchronously get platform properties new Thread ( ) { @ Override public void run ( ) { PlatformProperties . getPlatformProperties ( ) ; } } . start ( ) ; URL url = Resources . getResource ( ReportMaker . class , "template.html" ) ; String contents = Resources . toString ( url , Charsets . UTF_8 ) ; Cluster cluster = catalog . getClusters ( ) . get ( "cluster" ) ; assert ( cluster != null ) ; Database db = cluster . getDatabases ( ) . get ( "database" ) ; assert ( db != null ) ; String statsData = getStatsHTML ( db , minHeap , warnings ) ; contents = contents . replace ( "##STATS##" , statsData ) ; // generateProceduresTable needs to happen before generateSchemaTable // because some metadata used in the later is generated in the former String procData = generateProceduresTable ( db . getTables ( ) , db . getProcedures ( ) ) ; contents = contents . replace ( "##PROCS##" , procData ) ; String schemaData = generateSchemaTable ( db ) ; contents = contents . replace ( "##SCHEMA##" , schemaData ) ; DatabaseSizes sizes = CatalogSizing . getCatalogSizes ( db , DrRoleType . XDCR . value ( ) . equals ( cluster . getDrrole ( ) ) ) ; String sizeData = generateSizeTable ( sizes ) ; contents = contents . replace ( "##SIZES##" , sizeData ) ; String clusterConfig = generateClusterConfiguration ( isPro , hostCount , sitesPerHost , kfactor ) ; contents = contents . replace ( "##CLUSTERCONFIG##" , clusterConfig ) ; String sizeSummary = generateSizeSummary ( sizes ) ; contents = contents . replace ( "##SIZESUMMARY##" , sizeSummary ) ; String heapSummary = generateRecommendedServerSettings ( sizes ) ; contents = contents . replace ( "##RECOMMENDEDSERVERSETTINGS##" , heapSummary ) ; String platformData = PlatformProperties . getPlatformProperties ( ) . toHTML ( ) ; contents = contents . replace ( "##PLATFORM##" , platformData ) ; contents = contents . replace ( "##VERSION##" , VoltDB . instance ( ) . getVersionString ( ) ) ; contents = contents . replace ( "##DDL##" , escapeHtml4 ( autoGenDDL ) ) ; DateFormat df = new SimpleDateFormat ( "d MMM yyyy HH:mm:ss z" ) ; contents = contents . replace ( "##TIMESTAMP##" , df . format ( m_timestamp ) ) ; String msg = Encoder . hexEncode ( VoltDB . instance ( ) . getVersionString ( ) + "," + System . currentTimeMillis ( ) ) ; contents = contents . replace ( "get.py?a=KEY&" , String . format ( "get.py?a=%s&" , msg ) ) ; return contents ; } | Generate the HTML catalog report from a newly compiled VoltDB catalog | 709 | 13 |
155,476 | public static String liveReport ( ) { byte [ ] reportbytes = VoltDB . instance ( ) . getCatalogContext ( ) . getFileInJar ( VoltCompiler . CATLOG_REPORT ) ; String report = new String ( reportbytes , Charsets . UTF_8 ) ; // remove commented out code report = report . replace ( "<!--##RESOURCES" , "" ) ; report = report . replace ( "##RESOURCES-->" , "" ) ; // inject the cluster overview //String clusterStr = "<h4>System Overview</h4>\n<p>" + getLiveSystemOverview() + "</p><br/>\n"; //report = report.replace("<!--##CLUSTER##-->", clusterStr); // inject the running system platform properties PlatformProperties pp = PlatformProperties . getPlatformProperties ( ) ; String ppStr = "<h4>Cluster Platform</h4>\n<p>" + pp . toHTML ( ) + "</p><br/>\n" ; report = report . replace ( "<!--##PLATFORM2##-->" , ppStr ) ; // change the live/static var to live if ( VoltDB . instance ( ) . getConfig ( ) . m_isEnterprise ) { report = report . replace ( "&b=r&" , "&b=e&" ) ; } else { report = report . replace ( "&b=r&" , "&b=c&" ) ; } return report ; } | Find the pre - compild catalog report in the jarfile and modify it for use in the the built - in web portal . | 328 | 26 |
155,477 | private static boolean turnOffClientInterface ( ) { // we don't expect this to ever fail, but if it does, skip to dying immediately VoltDBInterface vdbInstance = instance ( ) ; if ( vdbInstance != null ) { ClientInterface ci = vdbInstance . getClientInterface ( ) ; if ( ci != null ) { if ( ! ci . ceaseAllPublicFacingTrafficImmediately ( ) ) { return false ; } } } return true ; } | turn off client interface as fast as possible | 101 | 8 |
155,478 | private static void sendCrashSNMPTrap ( String msg ) { if ( msg == null || msg . trim ( ) . isEmpty ( ) ) { return ; } VoltDBInterface vdbInstance = instance ( ) ; if ( vdbInstance == null ) { return ; } SnmpTrapSender snmp = vdbInstance . getSnmpTrapSender ( ) ; if ( snmp == null ) { return ; } try { snmp . crash ( msg ) ; } catch ( Throwable t ) { VoltLogger log = new VoltLogger ( "HOST" ) ; log . warn ( "failed to issue a crash SNMP trap" , t ) ; } } | send a SNMP trap crash notification | 146 | 7 |
155,479 | public static void crashGlobalVoltDB ( String errMsg , boolean stackTrace , Throwable t ) { // for test code wasCrashCalled = true ; crashMessage = errMsg ; if ( ignoreCrash ) { throw new AssertionError ( "Faux crash of VoltDB successful." ) ; } // end test code // send a snmp trap crash notification sendCrashSNMPTrap ( errMsg ) ; try { // turn off client interface as fast as possible // we don't expect this to ever fail, but if it does, skip to dying immediately if ( ! turnOffClientInterface ( ) ) { return ; // this will jump to the finally block and die faster } // instruct the rest of the cluster to die instance ( ) . getHostMessenger ( ) . sendPoisonPill ( errMsg ) ; // give the pill a chance to make it through the network buffer Thread . sleep ( 500 ) ; } catch ( Exception e ) { e . printStackTrace ( ) ; // sleep even on exception in case the pill got sent before the exception try { Thread . sleep ( 500 ) ; } catch ( InterruptedException e2 ) { } } // finally block does its best to ensure death, no matter what context this // is called in finally { crashLocalVoltDB ( errMsg , stackTrace , t ) ; } } | Exit the process with an error message optionally with a stack trace . Also notify all connected peers that the node is going down . | 283 | 25 |
155,480 | public static void main ( String [ ] args ) { //Thread.setDefaultUncaughtExceptionHandler(new VoltUncaughtExceptionHandler()); Configuration config = new Configuration ( args ) ; try { if ( ! config . validate ( ) ) { System . exit ( - 1 ) ; } else { if ( config . m_startAction == StartAction . GET ) { cli ( config ) ; } else { initialize ( config ) ; instance ( ) . run ( ) ; } } } catch ( OutOfMemoryError e ) { String errmsg = "VoltDB Main thread: ran out of Java memory. This node will shut down." ; VoltDB . crashLocalVoltDB ( errmsg , false , e ) ; } } | Entry point for the VoltDB server process . | 154 | 9 |
155,481 | public static String getDefaultReplicationInterface ( ) { if ( m_config . m_drInterface == null || m_config . m_drInterface . isEmpty ( ) ) { if ( m_config . m_externalInterface == null ) { return "" ; } else { return m_config . m_externalInterface ; } } else { return m_config . m_drInterface ; } } | Selects the a specified m_drInterface over a specified m_externalInterface from m_config | 85 | 20 |
155,482 | public void removeAllZeros ( ) { Iterator < Map . Entry < K , AtomicLong > > entryIterator = map . entrySet ( ) . iterator ( ) ; while ( entryIterator . hasNext ( ) ) { Map . Entry < K , AtomicLong > entry = entryIterator . next ( ) ; AtomicLong atomic = entry . getValue ( ) ; if ( atomic != null && atomic . get ( ) == 0L ) { entryIterator . remove ( ) ; } } } | Removes all mappings from this map whose values are zero . | 102 | 13 |
155,483 | public CompletableFuture < ClientResponseWithPartitionKey [ ] > callAllPartitionProcedure ( String procedureName , Object ... params ) { return m_runner . callAllPartitionProcedure ( procedureName , params ) ; } | A version of the similar API from VoltDB clients but for for non - transactional procedures . Runs a single - partition procedure on every partition that exists at the time it is called . | 52 | 37 |
155,484 | public static ByteBuffer getTableDataReference ( VoltTable vt ) { ByteBuffer buf = vt . m_buffer . duplicate ( ) ; buf . rewind ( ) ; return buf ; } | End users should not call this method . Obtain a reference to the table s underlying buffer . The returned reference s position and mark are independent of the table s buffer position and mark . The returned buffer has no mark and is at position 0 . | 41 | 49 |
155,485 | private long reserveNextTicket ( double requiredPermits , long nowMicros ) { resync ( nowMicros ) ; long microsToNextFreeTicket = Math . max ( 0 , nextFreeTicketMicros - nowMicros ) ; double storedPermitsToSpend = Math . min ( requiredPermits , this . storedPermits ) ; double freshPermits = requiredPermits - storedPermitsToSpend ; long waitMicros = storedPermitsToWaitTime ( this . storedPermits , storedPermitsToSpend ) + ( long ) ( freshPermits * stableIntervalMicros ) ; this . nextFreeTicketMicros = nextFreeTicketMicros + waitMicros ; this . storedPermits -= storedPermitsToSpend ; return microsToNextFreeTicket ; } | Reserves next ticket and returns the wait time that the caller must wait for . | 176 | 16 |
155,486 | public Options addOptionGroup ( OptionGroup group ) { if ( group . isRequired ( ) ) { requiredOpts . add ( group ) ; } for ( Option option : group . getOptions ( ) ) { // an Option cannot be required if it is in an // OptionGroup, either the group is required or // nothing is required option . setRequired ( false ) ; addOption ( option ) ; optionGroups . put ( option . getKey ( ) , group ) ; } return this ; } | Add the specified option group . | 104 | 6 |
155,487 | public Options addOption ( String opt , String description ) { addOption ( opt , null , false , description ) ; return this ; } | Add an option that only contains a short name . The option does not take an argument . | 28 | 18 |
155,488 | public Options addOption ( String opt , boolean hasArg , String description ) { addOption ( opt , null , hasArg , description ) ; return this ; } | Add an option that only contains a short - name . It may be specified as requiring an argument . | 33 | 20 |
155,489 | public Options addOption ( String opt , String longOpt , boolean hasArg , String description ) { addOption ( new Option ( opt , longOpt , hasArg , description ) ) ; return this ; } | Add an option that contains a short - name and a long - name . It may be specified as requiring an argument . | 42 | 24 |
155,490 | public static VoltXMLElement mergeTwoElementsUsingOperator ( String opName , String opElementId , VoltXMLElement first , VoltXMLElement second ) { if ( first == null || second == null ) { return first == null ? second : first ; } if ( opName == null || opElementId == null ) { return null ; } VoltXMLElement retval = new VoltXMLElement ( "operation" ) ; retval . attributes . put ( "id" , opElementId ) ; retval . attributes . put ( "optype" , opName ) ; retval . children . add ( first ) ; retval . children . add ( second ) ; return retval ; } | If one of the elements is null return the other one diectly . | 150 | 15 |
155,491 | public static List < VoltXMLElement > buildLimitElements ( int limit , String limitValueElementId ) { if ( limitValueElementId == null ) { return null ; } List < VoltXMLElement > retval = new ArrayList < VoltXMLElement > ( ) ; retval . add ( new VoltXMLElement ( "offset" ) ) ; VoltXMLElement limitElement = new VoltXMLElement ( "limit" ) ; String strLimit = String . valueOf ( limit ) ; limitElement . attributes . put ( "limit" , strLimit ) ; limitElement . children . add ( buildValueElement ( limitValueElementId , false , strLimit , "BIGINT" ) ) ; retval . add ( limitElement ) ; return retval ; } | Build VoltXMLElement for expression like LIMIT 1 . | 165 | 12 |
155,492 | public static VoltXMLElement buildColumnParamJoincondElement ( String opName , VoltXMLElement leftElement , String valueParamElementId , String opElementId ) { VoltXMLElement valueParamElement = buildValueElement ( valueParamElementId ) ; return mergeTwoElementsUsingOperator ( opName , opElementId , leftElement , valueParamElement ) ; } | Build VoltXMLElement for expression like column = ? . | 80 | 12 |
155,493 | public static VoltXMLElement buildParamElement ( String elementId , String index , String valueType ) { VoltXMLElement retval = new VoltXMLElement ( "parameter" ) ; retval . attributes . put ( "id" , elementId ) ; retval . attributes . put ( "index" , index ) ; retval . attributes . put ( "valuetype" , valueType ) ; return retval ; } | Build an element to be inserted under the parameters tree . | 93 | 11 |
155,494 | @ Override public void loadFromJSONObject ( JSONObject jobj , Database db ) throws JSONException { super . loadFromJSONObject ( jobj , db ) ; m_lookupType = IndexLookupType . get ( jobj . getString ( Members . LOOKUP_TYPE . name ( ) ) ) ; m_sortDirection = SortDirectionType . get ( jobj . getString ( Members . SORT_DIRECTION . name ( ) ) ) ; if ( jobj . has ( Members . HAS_OFFSET_RANK . name ( ) ) ) { m_hasOffsetRankOptimization = jobj . getBoolean ( Members . HAS_OFFSET_RANK . name ( ) ) ; } m_purpose = jobj . has ( Members . PURPOSE . name ( ) ) ? jobj . getInt ( Members . PURPOSE . name ( ) ) : FOR_SCANNING_PERFORMANCE_OR_ORDERING ; m_targetIndexName = jobj . getString ( Members . TARGET_INDEX_NAME . name ( ) ) ; m_catalogIndex = db . getTables ( ) . get ( super . m_targetTableName ) . getIndexes ( ) . get ( m_targetIndexName ) ; // load end_expression m_endExpression = AbstractExpression . fromJSONChild ( jobj , Members . END_EXPRESSION . name ( ) , m_tableScan ) ; // load initial_expression m_initialExpression = AbstractExpression . fromJSONChild ( jobj , Members . INITIAL_EXPRESSION . name ( ) , m_tableScan ) ; // load searchkey_expressions AbstractExpression . loadFromJSONArrayChild ( m_searchkeyExpressions , jobj , Members . SEARCHKEY_EXPRESSIONS . name ( ) , m_tableScan ) ; // load COMPARE_NOTDISTINCT flag vector loadBooleanArrayFromJSONObject ( jobj , Members . COMPARE_NOTDISTINCT . name ( ) , m_compareNotDistinct ) ; // load skip_null_predicate m_skip_null_predicate = AbstractExpression . fromJSONChild ( jobj , Members . SKIP_NULL_PREDICATE . name ( ) , m_tableScan ) ; } | all members loaded | 502 | 3 |
155,495 | public boolean isPredicatesOptimizableForAggregate ( ) { // for reverse scan, need to examine "added" predicates List < AbstractExpression > predicates = ExpressionUtil . uncombinePredicate ( m_predicate ) ; // if the size of predicates doesn't equal 1, can't be our added artifact predicates if ( predicates . size ( ) != 1 ) { return false ; } // examin the possible "added" predicates: NOT NULL expr. AbstractExpression expr = predicates . get ( 0 ) ; if ( expr . getExpressionType ( ) != ExpressionType . OPERATOR_NOT ) { return false ; } if ( expr . getLeft ( ) . getExpressionType ( ) != ExpressionType . OPERATOR_IS_NULL ) { return false ; } // Not reverse scan. if ( m_lookupType != IndexLookupType . LT && m_lookupType != IndexLookupType . LTE ) { return false ; } return true ; } | added for reverse scan purpose only | 213 | 6 |
155,496 | private void respondWithDummy ( ) { final FragmentResponseMessage response = new FragmentResponseMessage ( m_fragmentMsg , m_initiator . getHSId ( ) ) ; response . m_sourceHSId = m_initiator . getHSId ( ) ; response . setRecovering ( true ) ; response . setStatus ( FragmentResponseMessage . SUCCESS , null ) ; // Set the dependencies even if this is a dummy response. This site could be the master // on elastic join, so the fragment response message is actually going to the MPI. for ( int frag = 0 ; frag < m_fragmentMsg . getFragmentCount ( ) ; frag ++ ) { final int outputDepId = m_fragmentMsg . getOutputDepId ( frag ) ; response . addDependency ( new DependencyPair . BufferDependencyPair ( outputDepId , m_rawDummyResponse , 0 , m_rawDummyResponse . length ) ) ; } response . setRespBufferable ( m_respBufferable ) ; m_initiator . deliver ( response ) ; } | Respond with a dummy fragment response . | 244 | 8 |
155,497 | public FragmentResponseMessage processFragmentTask ( SiteProcedureConnection siteConnection ) { final FragmentResponseMessage currentFragResponse = new FragmentResponseMessage ( m_fragmentMsg , m_initiator . getHSId ( ) ) ; currentFragResponse . setStatus ( FragmentResponseMessage . SUCCESS , null ) ; for ( int frag = 0 ; frag < m_fragmentMsg . getFragmentCount ( ) ; frag ++ ) { final long fragmentId = VoltSystemProcedure . hashToFragId ( m_fragmentMsg . getPlanHash ( frag ) ) ; // equivalent to dep.depId: // final int outputDepId = m_fragmentMsg.getOutputDepId(frag); final VoltTrace . TraceEventBatch traceLog = VoltTrace . log ( VoltTrace . Category . SPSITE ) ; if ( traceLog != null ) { traceLog . add ( ( ) -> VoltTrace . beginDuration ( "runfragmenttask" , "txnId" , TxnEgo . txnIdToString ( getTxnId ( ) ) , "partition" , Integer . toString ( siteConnection . getCorrespondingPartitionId ( ) ) , "fragmentId" , String . valueOf ( fragmentId ) ) ) ; } ParameterSet params = m_fragmentMsg . getParameterSetForFragment ( frag ) ; try { // run the overloaded sysproc planfragment. pass an empty dependency // set since remote (non-aggregator) fragments don't receive dependencies. final DependencyPair dep = siteConnection . executeSysProcPlanFragment ( m_txnState , m_inputDeps , fragmentId , params ) ; // @Shutdown returns null, handle it here if ( dep != null ) { currentFragResponse . addDependency ( dep ) ; } } catch ( final EEException | SQLException | ReplicatedTableException e ) { hostLog . l7dlog ( Level . TRACE , LogKeys . host_ExecutionSite_ExceptionExecutingPF . name ( ) , new Object [ ] { Encoder . hexEncode ( m_fragmentMsg . getFragmentPlan ( frag ) ) } , e ) ; currentFragResponse . setStatus ( FragmentResponseMessage . UNEXPECTED_ERROR , e ) ; addDependencyToFragment ( currentFragResponse ) ; break ; } catch ( final SerializableException e ) { // Note that with SerializableException, the error code here might get changed before // the client/user sees it. It really just needs to indicate failure. // // Key point here vs the next catch block for VAE is to not wrap the subclass of // SerializableException here to preserve it during the serialization. // currentFragResponse . setStatus ( FragmentResponseMessage . USER_ERROR , e ) ; addDependencyToFragment ( currentFragResponse ) ; break ; } catch ( final VoltAbortException e ) { currentFragResponse . setStatus ( FragmentResponseMessage . USER_ERROR , new SerializableException ( CoreUtils . throwableToString ( e ) ) ) ; addDependencyToFragment ( currentFragResponse ) ; break ; } if ( traceLog != null ) { traceLog . add ( VoltTrace :: endDuration ) ; } } // we should never rollback DR buffer for MP sysprocs because we don't report the DR buffer size and therefore don't know if it is empty or not. currentFragResponse . setDrBufferSize ( 1 ) ; return currentFragResponse ; } | modified to work in the new world | 780 | 7 |
155,498 | @ Override public void run ( ) { try { VoltTable partitionKeys = null ; partitionKeys = m_client . callProcedure ( "@GetPartitionKeys" , "INTEGER" ) . getResults ( ) [ 0 ] ; while ( partitionKeys . advanceRow ( ) ) { m_client . callProcedure ( new NullCallback ( ) , "DeleteOldAdRequests" , partitionKeys . getLong ( "PARTITION_KEY" ) , m_expiredAgeInSeconds ) ; } m_client . callProcedure ( new NullCallback ( ) , "DeleteExpiredBids" ) ; } catch ( IOException | ProcCallException ex ) { ex . printStackTrace ( ) ; } } | Remove aged - out data from the ad_requests table . This table is partitioned and may be large so use the run - everywhere pattern to minimize impact to throughput . | 158 | 35 |
155,499 | public void updateLobUsage ( boolean commit ) { if ( ! hasLobOps ) { return ; } hasLobOps = false ; if ( commit ) { for ( int i = 0 ; i < createdLobs . size ( ) ; i ++ ) { long lobID = createdLobs . get ( i ) ; int delta = lobUsageCount . get ( lobID , 0 ) ; if ( delta == 1 ) { lobUsageCount . remove ( lobID ) ; createdLobs . remove ( i ) ; i -- ; } else if ( ! session . isBatch ) { database . lobManager . adjustUsageCount ( lobID , delta - 1 ) ; lobUsageCount . remove ( lobID ) ; createdLobs . remove ( i ) ; i -- ; } } if ( ! lobUsageCount . isEmpty ( ) ) { Iterator it = lobUsageCount . keySet ( ) . iterator ( ) ; while ( it . hasNext ( ) ) { long lobID = it . nextLong ( ) ; int delta = lobUsageCount . get ( lobID ) ; database . lobManager . adjustUsageCount ( lobID , delta - 1 ) ; } lobUsageCount . clear ( ) ; } return ; } else { for ( int i = 0 ; i < createdLobs . size ( ) ; i ++ ) { long lobID = createdLobs . get ( i ) ; database . lobManager . deleteLob ( lobID ) ; } createdLobs . clear ( ) ; lobUsageCount . clear ( ) ; return ; } } | update LobManager user counts delete lobs that have no usage | 330 | 12 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.