input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testGetConstraintTextWhenItemDefinitionHasAllowedValues ( ) { final org . kie . workbench . common . dmn . api . definition . v1_1 . ItemDefinition itemDefinition = mock ( org . kie . workbench . common . dmn . api . definition . v1_1 . ItemDefinition . class ) ; final org . kie . workbench . common . dmn . api . definition . v1_1 . UnaryTests allowedValues = mock ( org . kie . workbench . common . dmn . api . definition . v1_1 . UnaryTests . class ) ; final java . lang . String expectedText = "(1..10)" ; when ( allowedValues . getText ( ) ) . thenReturn ( new org . kie . workbench . common . dmn . api . property . dmn . Text ( expectedText ) ) ; when ( itemDefinition . getAllowedValues ( ) ) . thenReturn ( allowedValues ) ; final java . lang . String actualText = utils . getConstraintText ( itemDefinition ) ; "<AssertPlaceHolder>" ; } getConstraintText ( org . kie . workbench . common . dmn . api . definition . v1_1 . ItemDefinition ) { return java . util . Optional . ofNullable ( itemDefinition . getAllowedValues ( ) ) . map ( UnaryTests :: getText ) . orElse ( new org . kie . workbench . common . dmn . api . property . dmn . Text ( ) ) . getValue ( ) ; }
org . junit . Assert . assertEquals ( expectedText , actualText )
testEchoServer ( ) { final java . util . concurrent . atomic . AtomicReference < java . lang . String > result = new java . util . concurrent . atomic . AtomicReference ( ) ; final java . util . concurrent . CountDownLatch latch = new java . util . concurrent . CountDownLatch ( 1 ) ; java . net . SocketAddress serverSocket = new java . net . InetSocketAddress ( 0 ) ; final spullara . nio . channels . FutureServerSocketChannel fssc = new spullara . nio . channels . FutureServerSocketChannel ( ) . bind ( serverSocket ) ; java . util . function . Consumer < spullara . nio . channels . FutureSocketChannel > accepted = new java . util . function . Consumer < spullara . nio . channels . FutureSocketChannel > ( ) { public void accept ( spullara . nio . channels . FutureSocketChannel fsc ) { fssc . accept ( ) . thenAccept ( this ) ; java . nio . ByteBuffer bb = java . nio . ByteBuffer . allocate ( 1024 ) ; fsc . read ( bb ) . thenAccept ( ( length ) -> { bb . flip ( ) ; fsc . write ( bb ) ; fsc . close ( ) ; } ) ; } } ; fssc . accept ( ) . thenAccept ( accepted ) ; spullara . nio . channels . FutureSocketChannel fsc = new spullara . nio . channels . FutureSocketChannel ( ) ; java . net . SocketAddress clientSocket = new java . net . InetSocketAddress ( "localhost" , fssc . getLocalAddress ( ) . getPort ( ) ) ; fsc . connect ( clientSocket ) . thenAccept ( ( v ) -> { fsc . write ( java . nio . ByteBuffer . wrap ( "hello" . getBytes ( ) ) ) . thenAccept ( ( sent ) -> { java . nio . ByteBuffer readBuffer = java . nio . ByteBuffer . allocate ( sent ) ; fsc . read ( readBuffer ) . thenAccept ( ( recv ) -> { readBuffer . flip ( ) ; result . set ( new java . lang . String ( readBuffer . array ( ) ) ) ; latch . countDown ( ) ; } ) ; } ) ; } ) ; latch . await ( ) ; "<AssertPlaceHolder>" ; } get ( ) { if ( ! ( set ) ) { synchronized ( this ) { if ( ! ( set ) ) { try { value = callable . call ( ) ; set = true ; } catch ( java . lang . Exception e ) { throw new java . lang . RuntimeException ( "Lazy<sp>initialization<sp>failure" , e ) ; } } } } return value ; }
org . junit . Assert . assertEquals ( "hello" , result . get ( ) )
test13lookupResource ( ) { java . lang . String serviceName = "HDFS_1" ; org . apache . ranger . plugin . service . ResourceLookupContext context = new org . apache . ranger . plugin . service . ResourceLookupContext ( ) ; context . setResourceName ( serviceName ) ; context . setUserInput ( "HDFS" ) ; java . util . List < java . lang . String > list = serviceREST . lookupResource ( serviceName , context ) ; "<AssertPlaceHolder>" ; } lookupResource ( java . lang . String , org . apache . ranger . plugin . service . ResourceLookupContext ) { if ( org . apache . ranger . rest . ServiceREST . LOG . isDebugEnabled ( ) ) { org . apache . ranger . rest . ServiceREST . LOG . debug ( ( ( "==><sp>ServiceREST.lookupResource(" + serviceName ) + ")" ) ) ; } java . util . List < java . lang . String > ret = new java . util . ArrayList < java . lang . String > ( ) ; org . apache . ranger . plugin . util . RangerPerfTracer perf = null ; try { if ( org . apache . ranger . plugin . util . RangerPerfTracer . isPerfTraceEnabled ( org . apache . ranger . rest . ServiceREST . PERF_LOG ) ) { perf = org . apache . ranger . plugin . util . RangerPerfTracer . getPerfTracer ( org . apache . ranger . rest . ServiceREST . PERF_LOG , ( ( "ServiceREST.lookupResource(serviceName=" + serviceName ) + ")" ) ) ; } ret = serviceMgr . lookupResource ( serviceName , context , svcStore ) ; } catch ( javax . ws . rs . WebApplicationException excp ) { throw excp ; } catch ( java . lang . Throwable excp ) { org . apache . ranger . rest . ServiceREST . LOG . error ( ( ( ( ( "lookupResource(" + serviceName ) + ",<sp>" ) + context ) + ")<sp>failed" ) , excp ) ; throw restErrorUtil . createRESTException ( excp . getMessage ( ) ) ; } finally { org . apache . ranger . plugin . util . RangerPerfTracer . log ( perf ) ; } if ( org . apache . ranger . rest . ServiceREST . LOG . isDebugEnabled ( ) ) { org . apache . ranger . rest . ServiceREST . LOG . debug ( ( ( ( "<==<sp>ServiceREST.lookupResource(" + serviceName ) + "):<sp>" ) + ret ) ) ; } return ret ; }
org . junit . Assert . assertNotNull ( list )
findInteractionsTaxaInteractionIndexTargetTaxaOnlyTaxonIdPrefix ( ) { java . util . HashMap < java . lang . String , java . lang . String [ ] > params = new java . util . HashMap < java . lang . String , java . lang . String [ ] > ( ) { { put ( "somePrefix" 3 , new java . lang . String [ ] { "somePrefix" 1 } ) ; put ( "interactionType" , new java . lang . String [ ] { "preyedUponBy" } ) ; put ( "somePrefix" 2 , new java . lang . String [ ] { "somePrefix" } ) ; put ( "field" , new java . lang . String [ ] { "source_taxon_name" , "target_taxon_name" } ) ; } } ; org . eol . globi . util . CypherQuery query = buildInteractionQuery ( params , org . eol . globi . server . MULTI_TAXON_DISTINCT_BY_NAME_ONLY ) ; java . util . Map < java . lang . String , java . lang . String > expected = new java . util . HashMap < java . lang . String , java . lang . String > ( ) { { put ( "target_taxon_name" , "path:\\\"Arthropoda\\\"somePrefix" 0 ) ; put ( "source_taxon_prefix" , "somePrefix.*" ) ; put ( "target_taxon_prefix" , "somePrefix.*" ) ; } } ; "<AssertPlaceHolder>" ; } getParams ( ) { return params ; }
org . junit . Assert . assertThat ( query . getParams ( ) , org . hamcrest . core . Is . is ( expected ) )
getRedirectMpUrlHttps_OK ( ) { java . lang . String value = "https://thisisaurl/?mId=aaa" ; when ( confSvcMock . getVOConfigurationSetting ( eq ( ConfigurationKey . MP_ERROR_REDIRECT_HTTPS ) , anyString ( ) ) ) . thenReturn ( org . oscm . ui . filter . MarketplaceContextFilterTest . createSetting ( ConfigurationKey . MP_ERROR_REDIRECT_HTTPS , value ) ) ; java . lang . String result = mpCtxFilter . getRedirectMpUrlHttps ( confSvcMock ) ; "<AssertPlaceHolder>" ; } getRedirectMpUrlHttps ( org . oscm . ui . filter . ConfigurationService ) { java . lang . String mpRedirect = cs . getVOConfigurationSetting ( ConfigurationKey . MP_ERROR_REDIRECT_HTTPS , Configuration . GLOBAL_CONTEXT ) . getValue ( ) ; return mpRedirect ; }
org . junit . Assert . assertEquals ( value , result )
getCurrentSchedulingList_A$ ( ) { java . util . List < com . github . seratch . taskun . scheduler . crond . RawCrontabLine > actual = taskun . getCurrentRawCrontabLines ( ) ; "<AssertPlaceHolder>" ; } getCurrentRawCrontabLines ( ) { return cronInvocation . getCurrentRawCrontabLines ( ) ; }
org . junit . Assert . assertNotNull ( actual )
testDocumentWrapper ( ) { com . orientechnologies . orient . core . id . ORID orid = new com . orientechnologies . orient . core . id . ORecordId ( "#5:0" ) ; com . orientechnologies . orient . core . record . impl . ODocument adminDocument = orid . getRecord ( ) ; com . orientechnologies . orient . core . metadata . security . OUser admin = ru . ydn . wicket . wicketorientdb . MainUtilsTest . wicket . getTester ( ) . getMetadata ( ) . getSecurity ( ) . getUser ( "admin" ) ; ru . ydn . wicket . wicketorientdb . utils . DocumentWrapperTransformer < com . orientechnologies . orient . core . metadata . security . OUser > transformer = new ru . ydn . wicket . wicketorientdb . utils . DocumentWrapperTransformer < com . orientechnologies . orient . core . metadata . security . OUser > ( com . orientechnologies . orient . core . metadata . security . OUser . class ) ; "<AssertPlaceHolder>" ; } apply ( T ) { return comparableValue ( input , sortParam . getProperty ( ) ) ; }
org . junit . Assert . assertEquals ( admin , transformer . apply ( adminDocument ) )
testIsValidConfiguration_solutionIsNoValid_warIs ( ) { doReturn ( ( ( org . pentaho . test . platform . utils . TestResourceLocation . TEST_RESOURCES ) + "/notAfolder.txt" ) ) . when ( spyAppConfigProperties ) . getSolutionPath ( ) ; doReturn ( "test-res" ) . when ( spyAppConfigProperties ) . getWarPath ( ) ; "<AssertPlaceHolder>" ; } isValidConfiguration ( ) { boolean solutionPathValid = false ; boolean warPathValid = false ; java . io . File solutionPathFile = new java . io . File ( getSolutionPath ( ) ) ; if ( ( solutionPathFile != null ) && ( solutionPathFile . isDirectory ( ) ) ) { solutionPathValid = true ; } java . io . File warPathFile = new java . io . File ( getWarPath ( ) ) ; if ( ( warPathFile != null ) && ( warPathFile . isDirectory ( ) ) ) { warPathValid = true ; } return solutionPathValid && warPathValid ; }
org . junit . Assert . assertFalse ( spyAppConfigProperties . isValidConfiguration ( ) )
testAddSuccess ( ) { com . wealthfront . thompsonsampling . ObservedArmPerformance a1 = new com . wealthfront . thompsonsampling . ObservedArmPerformance ( 2 , 3 ) ; a1 . addSuccess ( ) ; "<AssertPlaceHolder>" ; } addSuccess ( ) { successes = ( successes ) + 1 ; return this ; }
org . junit . Assert . assertEquals ( new com . wealthfront . thompsonsampling . ObservedArmPerformance ( 3 , 3 ) , a1 )
testPredicate ( ) { int sum = java . util . stream . Stream . of ( 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 ) . filter ( new sirius . kernel . commons . Limit ( 2 , 4 ) . asPredicate ( ) ) . mapToInt ( ( i ) -> i ) . sum ( ) ; "<AssertPlaceHolder>" ; } asPredicate ( ) { return ( object ) -> nextRow ( ) ; }
org . junit . Assert . assertEquals ( ( ( ( 3 + 4 ) + 5 ) + 6 ) , sum )
testEmptyTextBoxesShouldReturnAnEmptyStringNotNull ( ) { driver . get ( pages . formPage ) ; org . openqa . selenium . WebElement emptyTextBox = driver . findElement ( org . openqa . selenium . By . id ( "working" ) ) ; "<AssertPlaceHolder>" ; } getAttribute ( java . lang . String ) { return commandProcessor . getString ( "getAttribute" , new java . lang . String [ ] { attributeLocator } ) ; }
org . junit . Assert . assertEquals ( emptyTextBox . getAttribute ( "value" ) , "" )
skipTest ( ) { org . apache . maven . shared . filtering . BoundedReader mr = new org . apache . maven . shared . filtering . BoundedReader ( sr , 3 ) ; mr . skip ( 2 ) ; mr . read ( ) ; "<AssertPlaceHolder>" ; mr . close ( ) ; } read ( ) { if ( ( replaceIndex ) > 0 ) { return replaceData . charAt ( ( ( replaceData . length ( ) ) - ( ( replaceIndex ) -- ) ) ) ; } if ( eof ) { return - 1 ; } org . apache . maven . shared . filtering . BoundedReader in = new org . apache . maven . shared . filtering . BoundedReader ( this . in , markLength ) ; int ch = in . read ( ) ; if ( ( ch == ( - 1 ) ) || ( ( ch == '\n' ) && ( ! ( supportMultiLineFiltering ) ) ) ) { return ch ; } boolean inEscape = ( useEscape ) && ( ch == ( getEscapeString ( ) . charAt ( 0 ) ) ) ; java . lang . StringBuilder key = new java . lang . StringBuilder ( ) ; if ( inEscape ) { for ( int i = 0 ; i < ( getEscapeString ( ) . length ( ) ) ; i ++ ) { key . append ( ( ( char ) ( ch ) ) ) ; if ( ( ( ch != ( getEscapeString ( ) . charAt ( i ) ) ) || ( ch == ( - 1 ) ) ) || ( ( ch == '\n' ) && ( ! ( supportMultiLineFiltering ) ) ) ) { in . reset ( ) ; inEscape = false ; key . setLength ( 0 ) ; break ; } ch = in . read ( ) ; } } int max = 0 ; for ( org . codehaus . plexus . interpolation . multi . DelimiterSpecification spec : delimiters ) { java . lang . String begin = spec . getBegin ( ) ; if ( ( begin . length ( ) ) < max ) { continue ; } for ( int i = 0 ; i < ( begin . length ( ) ) ; i ++ ) { if ( ( ( ch != ( begin . charAt ( i ) ) ) || ( ch == ( - 1 ) ) ) || ( ( ch == '\n' ) && ( ! ( supportMultiLineFiltering ) ) ) ) { break ; } if ( i == ( ( begin . length ( ) ) - 1 ) ) { beginToken = spec . getBegin ( ) ; endToken = spec . getEnd ( ) ; } ch = in . read ( ) ; } in . reset ( ) ; in . skip ( key . length ( ) ) ; ch = in . read ( ) ; } if ( inEscape ) { if ( ( beginToken ) != null ) { if ( ! ( isPreserveEscapeString ( ) ) ) { key . setLength ( 0 ) ; } } beginToken = null ; endToken = null ; key . append ( ( ( char ) ( ch ) ) ) ; replaceData = key . toString ( ) ; replaceIndex = key . length ( ) ; return read ( ) ; } if ( ( ( ( ( beginToken ) == null ) || ( ( beginToken . length ( ) ) == 0 ) ) || ( ( endToken ) == null ) ) || ( ( endToken . length ( ) ) == 0 ) ) { in . reset ( ) ; return in . read ( ) ; } key . append ( beginToken ) ; in . reset ( ) ; in . skip ( beginToken . length ( ) ) ; ch = in . read ( ) ; int endTokenSize = endToken . length ( ) ; int end = endTokenSize ; do { if ( ch == ( - 1 ) ) { break ; } else if ( ( ch == '\n' ) && ( ! ( supportMultiLineFiltering ) ) ) { key . append ( ( ( char ) ( ch ) ) ) ; break ; } key . append ( ( ( char ) ( ch ) ) ) ; if ( ch == ( this . endToken . charAt ( ( endTokenSize - end ) ) ) ) { end -- ; if ( end == 0 ) { break ; } } else { end = endTokenSize ; } ch = in . read ( ) ; } while ( true ) ; beginToken = null ; endToken = null ; java . lang . String value = null ; if ( end == 0 ) { try { if ( interpolateWithPrefixPattern ) { value = interpolator . interpolate ( key . toString ( ) , "" , recursionInterceptor ) ; } else { value = interpolator . interpolate ( key . toString ( ) , recursionInterceptor ) ; } } catch ( org . codehaus . plexus . interpolation . InterpolationException e ) { java . lang . IllegalArgumentException error = new java . lang . IllegalArgumentException ( e . getMessage ( ) ) ; error . initCause ( e ) ; throw error ; } } else { in . reset ( ) ; return in . read ( ) ; } if ( value != null ) { replaceData = value ; replaceIndex = value . length ( ) ; } else { replaceData = key . toString ( ) ; replaceIndex = key . length ( ) ; } if ( ch == ( - 1 ) ) {
org . junit . Assert . assertEquals ( ( - 1 ) , mr . read ( ) )
testAddAll ( ) { final org . scijava . object . ObjectIndex < java . lang . Object > objectIndex = new org . scijava . object . ObjectIndex ( java . lang . Object . class ) ; final java . util . ArrayList < java . lang . Object > objects = new java . util . ArrayList ( ) ; objects . add ( new java . lang . Integer ( 5 ) ) ; objects . add ( new java . lang . Float ( 2.5F ) ) ; objects . add ( new java . lang . Integer ( 3 ) ) ; objectIndex . addAll ( objects ) ; final java . util . List < java . lang . Object > result = objectIndex . getAll ( ) ; "<AssertPlaceHolder>" ; } getAll ( ) { return get ( org . scijava . object . ObjectIndex . All . class ) ; }
org . junit . Assert . assertEquals ( objects , result )
test_WithOnlyNameMismatch ( ) { ctx . request ( ) . addHeader ( "randomHeader" , null ) ; extension . activate ( configWithRequestHeader ) ; boolean actual = extension . accepts ( ctx . request ( ) , null , extension . getAllowedKeyValues ( ) ) ; "<AssertPlaceHolder>" ; } activate ( java . util . Map ) { this . listRoot = org . apache . sling . commons . osgi . PropertiesUtil . toString ( props . get ( com . adobe . acs . commons . genericlists . impl . GenericListJsonResourceProvider . PROP_LIST_ROOT ) , com . adobe . acs . commons . genericlists . impl . GenericListJsonResourceProvider . DEFAULT_LIST_ROOT ) ; }
org . junit . Assert . assertFalse ( actual )
testArrayLatLon ( ) { try ( ucar . nc2 . dt . RadialDatasetSweep ds = arrayLatLonData ( ) ) { "<AssertPlaceHolder>" ; } } isStationary ( ) { return true ; }
org . junit . Assert . assertTrue ( ds . isStationary ( ) )
testCreate ( ) { java . util . Map < java . lang . Object , java . lang . Object > properties = new java . util . HashMap < java . lang . Object , java . lang . Object > ( ) ; org . geotools . data . ws . WSDataStoreFactory dsf = new org . geotools . data . ws . WSDataStoreFactory ( ) ; properties . put ( "WSDataStoreFactory:GET_CONNECTION_URL" , "http://d00109:8080/xaware/XADocSoapServlet" ) ; properties . put ( "WSDataStoreFactory:TIMEOUT" , java . lang . Integer . valueOf ( 30000 ) ) ; properties . put ( "WSDataStoreFactory:TEMPLATE_DIRECTORY" , org . geotools . data . ws . WS_DataStoreTest . TEST_DIRECTORY ) ; properties . put ( "WSDataStoreFactory:TEMPLATE_NAME" , "request.ftl" ) ; properties . put ( "WSDataStoreFactory:CAPABILITIES_FILE_LOCATION" , ( ( org . geotools . data . ws . WS_DataStoreTest . TEST_DIRECTORY ) + "ws_capabilities_equals_removed.xml" ) ) ; org . geotools . data . ws . XmlDataStore ds = dsf . createDataStore ( properties ) ; "<AssertPlaceHolder>" ; } createDataStore ( java . util . Map ) { org . geotools . jdbc . JDBCDataStore dataStore = new org . geotools . jdbc . JDBCDataStore ( ) ; final org . geotools . jdbc . SQLDialect dialect = createSQLDialect ( dataStore ) ; dataStore . setSQLDialect ( dialect ) ; javax . sql . DataSource ds = ( ( javax . sql . DataSource ) ( org . geotools . jdbc . JDBCDataStoreFactory . DATASOURCE . lookUp ( params ) ) ) ; if ( ds != null ) { dataStore . setDataSource ( ds ) ; } else { dataStore . setDataSource ( createDataSource ( params , dialect ) ) ; } java . lang . Integer fetchSize = ( ( java . lang . Integer ) ( org . geotools . jdbc . JDBCDataStoreFactory . FETCHSIZE . lookUp ( params ) ) ) ; if ( ( fetchSize != null ) && ( fetchSize > 0 ) ) dataStore . setFetchSize ( fetchSize ) ; java . lang . Integer batchInsertSize = ( ( java . lang . Integer ) ( org . geotools . jdbc . JDBCDataStoreFactory . BATCH_INSERT_SIZE . lookUp ( params ) ) ) ; if ( ( batchInsertSize != null ) && ( batchInsertSize > 0 ) ) { dataStore . setBatchInsertSize ( batchInsertSize ) ; } java . lang . String namespace = ( ( java . lang . String ) ( org . geotools . jdbc . JDBCDataStoreFactory . NAMESPACE . lookUp ( params ) ) ) ; if ( namespace != null ) { dataStore . setNamespaceURI ( namespace ) ; } java . lang . String schema = ( ( java . lang . String ) ( org . geotools . jdbc . JDBCDataStoreFactory . SCHEMA . lookUp ( params ) ) ) ; if ( schema != null ) { dataStore . setDatabaseSchema ( schema ) ; } java . lang . String metadataTable = ( ( java . lang . String ) ( org . geotools . jdbc . JDBCDataStoreFactory . PK_METADATA_TABLE . lookUp ( params ) ) ) ; if ( metadataTable != null ) { org . geotools . jdbc . MetadataTablePrimaryKeyFinder tableFinder = new org . geotools . jdbc . MetadataTablePrimaryKeyFinder ( ) ; if ( metadataTable . contains ( "." ) ) { java . lang . String [ ] parts = metadataTable . split ( "\\." ) ; if ( ( parts . length ) > 2 ) throw new java . lang . IllegalArgumentException ( ( "The<sp>primary<sp>key<sp>metadata<sp>table<sp>format<sp>" + "is<sp>either<sp>'name'<sp>or<sp>'schema.name'" ) ) ; tableFinder . setTableSchema ( parts [ 0 ] ) ; tableFinder . setTableName ( parts [ 1 ] ) ; } else { tableFinder . setTableName ( metadataTable ) ; } dataStore . setPrimaryKeyFinder ( new org . geotools . jdbc . CompositePrimaryKeyFinder ( tableFinder , new org . geotools . jdbc . HeuristicPrimaryKeyFinder ( ) ) ) ; } java . lang . Boolean exposePk = ( ( java . lang . Boolean ) ( org . geotools . jdbc . JDBCDataStoreFactory . EXPOSE_PK . lookUp ( params ) ) ) ; if ( exposePk != null ) { dataStore . setExposePrimaryKeyColumns ( exposePk ) ; } java . lang . String sqlOnBorrow = ( ( java . lang . String ) ( org . geotools . jdbc . JDBCDataStoreFactory . SQL_ON_BORROW . lookUp ( params ) ) ) ; java . lang . String sqlOnRelease = ( ( java . lang . String ) ( org . geotools . jdbc . JDBCDataStoreFactory . SQL_ON_RELEASE . lookUp ( params ) ) ) ; if ( ( sqlOnBorrow != null ) || ( sqlOnRelease != null ) ) { org . geotools . jdbc . SessionCommandsListener listener = new org . geotools . jdbc . SessionCommandsListener ( sqlOnBorrow , sqlOnRelease ) ; dataStore . getConnectionLifecycleListeners ( ) . add ( listener ) ; } java . lang . String callbackFactory = ( ( java . lang . String ) ( org . geotools . jdbc . JDBCDataStoreFactory . CALLBACK_FACTORY . lookUp ( params ) ) ) ; if ( callbackFactory == null ) { callbackFactory = java . lang . System . getProperty ( org . geotools . jdbc . JDBCDataStoreFactory . CALLBACK_PROPERTY ) ; } dataStore . setCallbackFactory ( ( callbackFactory != null ? org . geotools . jdbc . JDBCDataStoreFactory . findCallbackFactory ( callbackFactory ) : JDBCCallbackFactory . NULL ) ) ; dataStore . setFilterFactory ( org . geotools . factory . CommonFactoryFinder . getFilterFactory ( null ) ) ; dataStore . setGeometryFactory ( new org . locationtech . jts . geom . GeometryFactory ( ) ) ; dataStore . setFeatureTypeFactory ( new org . geotools . feature . type . FeatureTypeFactoryImpl ( ) ) ; dataStore . setFeatureFactory ( org . geotools . factory . CommonFactoryFinder . getFeatureFactory ( null ) ) ; dataStore . setDataStoreFactory ( this ) ; org . geotools . jdbc
org . junit . Assert . assertNotNull ( ds )
shouldReturnNullIfClassDissapearsWhileParsing ( ) { org . infinitest . parser . JavaAssistClassParser mockParser = mock ( org . infinitest . parser . JavaAssistClassParser . class ) ; org . infinitest . parser . JavaClassBuilder builder = new org . infinitest . parser . JavaClassBuilder ( mockParser ) ; org . infinitest . parser . NotFoundException cause = new org . infinitest . parser . NotFoundException ( "" ) ; when ( mockParser . classFileChanged ( new org . infinitest . parser . File ( "UNKNOWN" ) ) ) . thenThrow ( new java . lang . RuntimeException ( cause ) ) ; "<AssertPlaceHolder>" ; } classFileChanged ( org . infinitest . parser . File ) { java . lang . String sha1 = org . infinitest . parser . Files . hash ( file , com . google . common . hash . Hashing . sha1 ( ) ) . toString ( ) ; org . infinitest . parser . JavaAssistClassParser . CacheEntry entry = org . infinitest . parser . JavaAssistClassParser . BY_PATH . get ( file . getAbsolutePath ( ) ) ; if ( ( entry != null ) && ( entry . sha1 . equals ( sha1 ) ) ) { return entry . classname ; } org . infinitest . parser . FileInputStream inputStream = null ; try { inputStream = new org . infinitest . parser . FileInputStream ( file ) ; org . infinitest . parser . CtClass ctClass = getClassPool ( ) . makeClass ( inputStream ) ; java . lang . String classname = ctClass . getName ( ) ; org . infinitest . parser . JavaAssistClassParser . CLASSES_BY_NAME . remove ( classname ) ; org . infinitest . parser . JavaAssistClassParser . BY_PATH . put ( file . getAbsolutePath ( ) , new org . infinitest . parser . JavaAssistClassParser . CacheEntry ( sha1 , classname ) ) ; return classname ; } finally { if ( inputStream != null ) { inputStream . close ( ) ; } } }
org . junit . Assert . assertNull ( builder . classFileChanged ( new org . infinitest . parser . File ( "UNKNOWN" ) ) )
getSecondaryGroupUserIDsWhenNoGroupsResultsReturnsEmptyList ( ) { groups = "" ; when ( result . getString ( configuration . webappSecondaryGroupGroupIDColumn ) ) . thenReturn ( groups ) ; java . util . List < java . lang . String > secondaryGroups = webGroupDao . getSecondaryGroupUserIDs ( GROUP_ID ) ; "<AssertPlaceHolder>" ; } getSecondaryGroupUserIDs ( java . lang . String ) { java . util . List < java . lang . String > userIDs = new java . util . ArrayList < java . lang . String > ( ) ; result = sql . sqlQuery ( getSecondaryGroupUserIDsReadQuery ( ) ) ; while ( result . next ( ) ) { java . util . List < java . lang . String > groupIDs = convertDelimitedIDString ( result . getString ( configuration . webappSecondaryGroupGroupIDColumn ) ) ; if ( groupIDs . contains ( groupID ) ) { userIDs . add ( result . getString ( configuration . webappSecondaryGroupUserIDColumn ) ) ; } } return userIDs ; }
org . junit . Assert . assertEquals ( 0 , secondaryGroups . size ( ) )
testConstructor ( ) { org . biojava . nbio . genome . io . fastq . Fastq fastq = new org . biojava . nbio . genome . io . fastq . Fastq ( "description" , "sequence" , "quality_" , FastqVariant . FASTQ_SANGER ) ; "<AssertPlaceHolder>" ; try { new org . biojava . nbio . genome . io . fastq . Fastq ( null , "sequence" , "quality_" , FastqVariant . FASTQ_SANGER ) ; org . junit . Assert . fail ( "ctr(null<sp>description)<sp>expected<sp>IllegalArgumentException" ) ; } catch ( java . lang . IllegalArgumentException e ) { } try { new org . biojava . nbio . genome . io . fastq . Fastq ( "description" , null , "quality_" , FastqVariant . FASTQ_SANGER ) ; org . junit . Assert . fail ( "ctr(null<sp>sequence)<sp>expected<sp>IllegalArgumentException" ) ; } catch ( java . lang . IllegalArgumentException e ) { } try { new org . biojava . nbio . genome . io . fastq . Fastq ( "description" , "sequence" , null , FastqVariant . FASTQ_SANGER ) ; org . junit . Assert . fail ( "ctr(null<sp>quality)<sp>expected<sp>IllegalArgumentException" ) ; } catch ( java . lang . IllegalArgumentException e ) { } try { new org . biojava . nbio . genome . io . fastq . Fastq ( "description" , "sequence" , "quality_" , null ) ; org . junit . Assert . fail ( "ctr(null<sp>variant)<sp>expected<sp>IllegalArgumentException" ) ; } catch ( java . lang . IllegalArgumentException e ) { } }
org . junit . Assert . assertNotNull ( fastq )
testIsMineForRepositoryWithoutAccountAndCommentWithAuthor ( ) { org . eclipse . mylyn . reviews . core . model . IUser user = org . eclipse . mylyn . reviews . core . model . CommentContainerTest . createUser ( "foo@bar.com" ) ; org . eclipse . mylyn . reviews . core . model . IRepository repo = org . eclipse . mylyn . reviews . core . model . CommentContainerTest . createRepository ( ) ; org . eclipse . mylyn . reviews . core . model . IComment comment = org . eclipse . mylyn . reviews . core . model . CommentContainerTest . createReviewWithCommentAndAuthor ( repo , user ) ; "<AssertPlaceHolder>" ; } isMine ( ) { org . eclipse . mylyn . reviews . core . model . IRepository repos = ( ( getReview ( ) ) != null ) ? getReview ( ) . getRepository ( ) : null ; return ( ( ( getAuthor ( ) ) != null ) && ( repos != null ) ) && ( ( ( repos . getAccount ( ) ) == ( getAuthor ( ) ) ) || ( ( ( ( repos . getAccount ( ) ) != null ) && ( ( repos . getAccount ( ) . getEmail ( ) ) != null ) ) && ( repos . getAccount ( ) . getEmail ( ) . equals ( getAuthor ( ) . getEmail ( ) ) ) ) ) ; }
org . junit . Assert . assertThat ( comment . isMine ( ) , org . hamcrest . Matchers . is ( false ) )
testQuarantineMissingRegionDir ( ) { java . lang . String table = name . getMethodName ( ) ; java . util . concurrent . ExecutorService exec = new java . util . concurrent . ScheduledThreadPoolExecutor ( 10 ) ; final org . apache . hadoop . fs . FileSystem fs = org . apache . hadoop . fs . FileSystem . get ( org . apache . hadoop . hbase . util . TestHBaseFsck . conf ) ; org . apache . hadoop . hbase . util . HBaseFsck hbck = new org . apache . hadoop . hbase . util . HBaseFsck ( org . apache . hadoop . hbase . util . TestHBaseFsck . conf , exec ) { public org . apache . hadoop . hbase . util . hbck . HFileCorruptionChecker createHFileCorruptionChecker ( boolean sidelineCorruptHFiles ) throws java . io . IOException { return new org . apache . hadoop . hbase . util . hbck . HFileCorruptionChecker ( org . apache . hadoop . hbase . util . TestHBaseFsck . conf , executor , sidelineCorruptHFiles ) { boolean attemptedFirstRegionDir = false ; protected void checkRegionDir ( org . apache . hadoop . fs . Path p ) throws java . io . IOException { if ( ! ( attemptedFirstRegionDir ) ) { attemptedFirstRegionDir = true ; "<AssertPlaceHolder>" ; } super . checkRegionDir ( p ) ; } } ; } } ; doQuarantineTest ( table , hbck , 3 , 0 , 0 , 0 , 1 ) ; } delete ( org . apache . hadoop . hbase . rest . client . Cluster , java . lang . String ) { org . apache . commons . httpclient . methods . DeleteMethod method = new org . apache . commons . httpclient . methods . DeleteMethod ( ) ; try { int code = execute ( cluster , method , null , path ) ; org . apache . commons . httpclient . Header [ ] headers = method . getResponseHeaders ( ) ; byte [ ] content = method . getResponseBody ( ) ; return new org . apache . hadoop . hbase . rest . client . Response ( code , headers , content ) ; } finally { method . releaseConnection ( ) ; } }
org . junit . Assert . assertTrue ( fs . delete ( p , true ) )
testSerialization ( ) { org . jfree . data . xy . XIntervalDataItem item1 = new org . jfree . data . xy . XIntervalDataItem ( 1.0 , 2.0 , 3.0 , 4.0 ) ; java . io . ByteArrayOutputStream buffer = new java . io . ByteArrayOutputStream ( ) ; java . io . ObjectOutput out = new java . io . ObjectOutputStream ( buffer ) ; out . writeObject ( item1 ) ; out . close ( ) ; java . io . ObjectInput in = new java . io . ObjectInputStream ( new java . io . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; org . jfree . data . xy . XIntervalDataItem item2 = ( ( org . jfree . data . xy . XIntervalDataItem ) ( in . readObject ( ) ) ) ; in . close ( ) ; "<AssertPlaceHolder>" ; } close ( ) { try { this . connection . close ( ) ; } catch ( java . lang . Exception e ) { System . err . println ( "JdbcXYDataset:<sp>swallowing<sp>exception." ) ; } }
org . junit . Assert . assertEquals ( item1 , item2 )
testDecimalMax ( ) { boolean expResult = true ; javax . xml . xpath . XPathExpression expr = xpath . compile ( "//xs:element[@name='decimalMaxField']//xs:maxInclusive/@value<sp>=<sp>104" ) ; boolean result = ( ( java . lang . Boolean ) ( expr . evaluate ( schemaDocument , XPathConstants . BOOLEAN ) ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( expResult , result )
parseWhenSendFatalErrorState ( ) { org . xwiki . mail . MailState state = org . xwiki . mail . MailState . parse ( "send_fatal_error" ) ; "<AssertPlaceHolder>" ; } parse ( java . lang . String ) { org . xwiki . mail . MailState result ; if ( ( state . equalsIgnoreCase ( org . xwiki . mail . MailState . PREPARE_SUCCESS . toString ( ) ) ) || ( state . equalsIgnoreCase ( "READY" ) ) ) { result = org . xwiki . mail . MailState . PREPARE_SUCCESS ; } else if ( state . equalsIgnoreCase ( org . xwiki . mail . MailState . PREPARE_ERROR . toString ( ) ) ) { result = org . xwiki . mail . MailState . PREPARE_ERROR ; } else if ( ( state . equalsIgnoreCase ( org . xwiki . mail . MailState . SEND_SUCCESS . toString ( ) ) ) || ( state . equalsIgnoreCase ( "SENT" ) ) ) { result = org . xwiki . mail . MailState . SEND_SUCCESS ; } else if ( ( state . equalsIgnoreCase ( org . xwiki . mail . MailState . SEND_ERROR . toString ( ) ) ) || ( state . equalsIgnoreCase ( "FAILED" ) ) ) { result = org . xwiki . mail . MailState . SEND_ERROR ; } else if ( state . equalsIgnoreCase ( org . xwiki . mail . MailState . SEND_FATAL_ERROR . toString ( ) ) ) { result = org . xwiki . mail . MailState . SEND_FATAL_ERROR ; } else { throw new java . lang . IllegalArgumentException ( java . lang . String . format ( "Invalid<sp>mail<sp>state<sp>[%s]" , state ) ) ; } return result ; }
org . junit . Assert . assertEquals ( MailState . SEND_FATAL_ERROR , state )
testAddition ( ) { org . hipparchus . analysis . polynomials . PolynomialFunction p1 = new org . hipparchus . analysis . polynomials . PolynomialFunction ( new double [ ] { - 2 , 1 } ) ; org . hipparchus . analysis . polynomials . PolynomialFunction p2 = new org . hipparchus . analysis . polynomials . PolynomialFunction ( new double [ ] { 2 , - 1 , 0 } ) ; checkNullPolynomial ( p1 . add ( p2 ) ) ; p2 = p1 . add ( p1 ) ; checkPolynomial ( p2 , "-4<sp>+<sp>2<sp>x" ) ; p1 = new org . hipparchus . analysis . polynomials . PolynomialFunction ( new double [ ] { 1 , - 4 , 2 } ) ; p2 = new org . hipparchus . analysis . polynomials . PolynomialFunction ( new double [ ] { - 1 , 3 , - 2 } ) ; p1 = p1 . add ( p2 ) ; "<AssertPlaceHolder>" ; checkPolynomial ( p1 , "-x" ) ; } degree ( ) { return ( coefficients . length ) - 1 ; }
org . junit . Assert . assertEquals ( 1 , p1 . degree ( ) )
basicFunctionTest4 ( ) { java . util . List < com . navercorp . pinpoint . profiler . receiver . service . WaitingJob > waitingJobList = this . waitingJobListFactory . createList ( com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpServiceTest . CREATE_SIZE , com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpServiceTest . JOB_TIMEOUT ) ; java . util . List < com . navercorp . pinpoint . profiler . context . active . ActiveTraceSnapshot > activeTraceInfoList = createMockActiveTraceInfoList ( com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpServiceTest . CREATE_SIZE , com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpServiceTest . DEFAULT_TIME_MILLIS , com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpServiceTest . TIME_DIFF_INTERVAL , waitingJobList ) ; java . util . List < com . navercorp . pinpoint . profiler . context . active . ActiveTraceSnapshot > activeTraceSnapshotList = shuffle ( activeTraceInfoList ) ; int targetThreadNameSize = 3 ; java . util . List < java . lang . String > threadNameList = extractThreadNameList ( activeTraceSnapshotList . subList ( 0 , targetThreadNameSize ) , targetThreadNameSize ) ; int targetTraceIdSize = 3 ; java . util . List < java . lang . Long > localTraceIdList = extractLocalTraceIdList ( activeTraceSnapshotList . subList ( targetThreadNameSize , com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpServiceTest . CREATE_SIZE ) , targetTraceIdSize ) ; com . navercorp . pinpoint . thrift . dto . command . TCmdActiveThreadLightDump tCmdActiveThreadDump = createRequest ( 0 , threadNameList , localTraceIdList ) ; com . navercorp . pinpoint . profiler . receiver . service . ActiveThreadLightDumpService service = createService ( activeTraceInfoList ) ; com . navercorp . pinpoint . thrift . dto . command . TCmdActiveThreadLightDumpRes response = ( ( com . navercorp . pinpoint . thrift . dto . command . TCmdActiveThreadLightDumpRes ) ( service . requestCommandService ( tCmdActiveThreadDump ) ) ) ; "<AssertPlaceHolder>" ; } getThreadDumpsSize ( ) { return ( this . threadDumps ) == null ? 0 : this . threadDumps . size ( ) ; }
org . junit . Assert . assertEquals ( ( targetThreadNameSize + targetTraceIdSize ) , response . getThreadDumpsSize ( ) )
available_onCachedStream ( ) { final java . lang . String testString = "helloWorld" ; final byte [ ] testData = testString . getBytes ( ) ; final java . io . InputStream is = new org . exist . util . io . FastByteArrayInputStream ( testData ) ; final org . exist . util . io . CachingFilterInputStream cfis = new org . exist . util . io . CachingFilterInputStream ( getNewCache ( is ) ) ; cfis . mark ( Integer . MAX_VALUE ) ; for ( int i = 0 ; i < ( testData . length ) ; i ++ ) { cfis . read ( ) ; } cfis . reset ( ) ; "<AssertPlaceHolder>" ; } available ( ) { return bis . available ( ) ; }
org . junit . Assert . assertEquals ( testData . length , cfis . available ( ) )
testValidateWithNullSrcPort ( ) { org . o3project . odenos . core . component . network . topology . Link link = links . get ( "link1_id123" ) ; org . mockito . Mockito . when ( nodes . get ( link . getSrcNode ( ) ) . getPort ( link . getSrcPort ( ) ) ) . thenReturn ( null ) ; target = org . mockito . Mockito . spy ( new org . o3project . odenos . core . component . network . topology . Topology ( "101" , nodes , links ) ) ; "<AssertPlaceHolder>" ; } validate ( ) { try { org . apache . commons . lang . Validate . notEmpty ( this . lowerNw , "lower_nw<sp>is<sp>empty" ) ; org . apache . commons . lang . Validate . notEmpty ( this . lowerNwNode , "lower_nw_node<sp>is<sp>empty" ) ; org . apache . commons . lang . Validate . notEmpty ( this . lowerNwPort , "lower_nw_port<sp>is<sp>empty" ) ; org . apache . commons . lang . Validate . notEmpty ( this . upperNw , "upper_nw<sp>is<sp>empty" ) ; org . apache . commons . lang . Validate . notEmpty ( this . upperNwNode , "upper_nw_node<sp>is<sp>empty" ) ; org . apache . commons . lang . Validate . notEmpty ( this . upperNwPort , "upper_nw_port<sp>is<sp>empty" ) ; return true ; } catch ( java . lang . IllegalArgumentException ex ) { org . o3project . odenos . component . linklayerizer . LinkLayerizerBoundary . log . warn ( ex . getMessage ( ) , ex ) ; return false ; } }
org . junit . Assert . assertFalse ( target . validate ( ) )
testGetJobDefinition ( ) { java . lang . String namespace = NAMESPACE ; java . lang . String jobName = JOB_NAME ; org . finra . herd . model . api . xml . JobDefinition jobDefinition = getJobDefinition ( org . finra . herd . rest . NAMESPACE , org . finra . herd . rest . JOB_NAME ) ; when ( jobDefinitionService . getJobDefinition ( namespace , jobName ) ) . thenReturn ( jobDefinition ) ; org . finra . herd . model . api . xml . JobDefinition resultJobDefinition = jobDefinitionRestController . getJobDefinition ( namespace , jobName ) ; verify ( jobDefinitionService ) . getJobDefinition ( namespace , jobName ) ; verifyNoMoreInteractions ( jobDefinitionService ) ; "<AssertPlaceHolder>" ; } getJobDefinition ( java . lang . String , java . lang . String ) { java . lang . String namespaceLocal = alternateKeyHelper . validateStringParameter ( "namespace" , namespace ) ; java . lang . String jobNameLocal = alternateKeyHelper . validateStringParameter ( "job<sp>name" , jobName ) ; org . finra . herd . model . jpa . JobDefinitionEntity jobDefinitionEntity = jobDefinitionDaoHelper . getJobDefinitionEntity ( namespaceLocal , jobNameLocal ) ; return createJobDefinitionFromEntity ( jobDefinitionEntity ) ; }
org . junit . Assert . assertEquals ( jobDefinition , resultJobDefinition )
testReloadWithOptions ( ) { initializeExpectedBucket ( 5 ) ; com . google . cloud . storage . BucketInfo updatedInfo = com . google . cloud . storage . BucketTest . BUCKET_INFO . toBuilder ( ) . setNotFoundPage ( "p" ) . build ( ) ; com . google . cloud . storage . Bucket expectedUpdatedBucket = new com . google . cloud . storage . Bucket ( serviceMockReturnsOptions , new com . google . cloud . storage . BucketInfo . BuilderImpl ( updatedInfo ) ) ; expect ( storage . getOptions ( ) ) . andReturn ( mockOptions ) ; expect ( storage . get ( updatedInfo . getName ( ) , Storage . BucketGetOption . metagenerationMatch ( 42L ) ) ) . andReturn ( expectedUpdatedBucket ) ; replay ( storage ) ; initializeBucket ( ) ; com . google . cloud . storage . Bucket updatedBucket = bucket . reload ( Bucket . BucketSourceOption . metagenerationMatch ( ) ) ; "<AssertPlaceHolder>" ; } metagenerationMatch ( ) { return new com . google . cloud . storage . Bucket . BucketSourceOption ( StorageRpc . Option . IF_METAGENERATION_MATCH ) ; }
org . junit . Assert . assertEquals ( expectedUpdatedBucket , updatedBucket )
testFindMyApplicationForm ( ) { securityContextManager . setBaz ( ) ; org . springframework . data . domain . Page < com . github . emailtohl . building . site . entities . flow . ApplicationForm > page = applicationFormService . findMyApplicationForm ( pageable ) ; "<AssertPlaceHolder>" ; } getTotalElements ( ) { return totalElements ; }
org . junit . Assert . assertTrue ( ( ( page . getTotalElements ( ) ) > 0 ) )
testSendCommandClosedAfterConnected ( ) { java . lang . String testResponse = "testResponse" ; createValidSocket ( ) ; java . util . List < java . lang . String > responseList = new java . util . ArrayList ( ) ; responseList . add ( testResponse ) ; when ( mockedBufferedReader . readLine ( ) ) . thenReturn ( "OK" ) . thenReturn ( testResponse ) . thenReturn ( null ) ; when ( mockSocket . isClosed ( ) ) . thenReturn ( true ) ; org . bff . javampd . command . MPDCommand command = new org . bff . javampd . command . MPDCommand ( "command" ) ; java . util . List < java . lang . String > response = new java . util . ArrayList ( socket . sendCommand ( command ) ) ; "<AssertPlaceHolder>" ; } sendCommand ( org . bff . javampd . command . MPDCommand ) { try { checkSocket ( ) ; return new java . util . ArrayList ( mpdSocket . sendCommand ( command ) ) ; } catch ( org . bff . javampd . command . MPDSecurityException se ) { org . bff . javampd . command . MPDCommandExecutor . LOGGER . warn ( "Connection<sp>exception<sp>while<sp>sending<sp>command<sp>{},<sp>will<sp>retry" , command . getCommand ( ) , se ) ; authenticate ( ) ; return new java . util . ArrayList ( mpdSocket . sendCommand ( command ) ) ; } }
org . junit . Assert . assertEquals ( testResponse , response . get ( 0 ) )
testFindContainer ( ) { final ch . cyberduck . core . Host host = new ch . cyberduck . core . Host ( new ch . cyberduck . core . openstack . SwiftProtocol ( ) , "identity.api.rackspacecloud.com" , new ch . cyberduck . core . Credentials ( java . lang . System . getProperties ( ) . getProperty ( "rackspace.key" ) , java . lang . System . getProperties ( ) . getProperty ( "rackspace.secret" ) ) ) ; final ch . cyberduck . core . openstack . SwiftSession session = new ch . cyberduck . core . openstack . SwiftSession ( host ) ; new ch . cyberduck . core . LoginConnectionService ( new ch . cyberduck . core . DisabledLoginCallback ( ) , new ch . cyberduck . core . DisabledHostKeyCallback ( ) , new ch . cyberduck . core . DisabledPasswordStore ( ) , new ch . cyberduck . core . DisabledProgressListener ( ) ) . connect ( session , ch . cyberduck . core . PathCache . empty ( ) , new ch . cyberduck . core . DisabledCancelCallback ( ) ) ; final ch . cyberduck . core . Path container = new ch . cyberduck . core . Path ( "test-iad-cyberduck" , java . util . EnumSet . of ( Path . Type . directory , Path . Type . volume ) ) ; container . attributes ( ) . setRegion ( "IAD" ) ; final ch . cyberduck . core . PathAttributes attributes = new ch . cyberduck . core . openstack . SwiftAttributesFinderFeature ( session ) . find ( container ) ; "<AssertPlaceHolder>" ; session . close ( ) ; } getType ( ) { return Type . dracoon ; }
org . junit . Assert . assertEquals ( java . util . EnumSet . of ( Path . Type . volume , Path . Type . directory ) , container . getType ( ) )
whenSavePersistent_thenNothingHappens ( ) { com . baeldung . persistence . model . Person person = new com . baeldung . persistence . model . Person ( ) ; person . setName ( "John" ) ; java . lang . Long id1 = ( ( java . lang . Long ) ( session . save ( person ) ) ) ; java . lang . Long id2 = ( ( java . lang . Long ) ( session . save ( person ) ) ) ; "<AssertPlaceHolder>" ; } save ( org . baeldung . demo . model . Foo ) { emf . unwrap ( org . hibernate . SessionFactory . class ) . getCurrentSession ( ) . saveOrUpdate ( foo ) ; }
org . junit . Assert . assertEquals ( id1 , id2 )
testQueryConfigWithApolloConfigNotModified ( ) { java . lang . String someClientSideReleaseKey = "1" ; java . lang . String someServerSideReleaseKey = someClientSideReleaseKey ; javax . servlet . http . HttpServletResponse someResponse = mock ( javax . servlet . http . HttpServletResponse . class ) ; when ( configService . loadConfig ( someAppId , someClientIp , someAppId , someClusterName , defaultNamespaceName , someDataCenter , someNotificationMessages ) ) . thenReturn ( someRelease ) ; when ( someRelease . getReleaseKey ( ) ) . thenReturn ( someServerSideReleaseKey ) ; com . ctrip . framework . apollo . core . dto . ApolloConfig result = configController . queryConfig ( someAppId , someClusterName , defaultNamespaceName , someDataCenter , java . lang . String . valueOf ( someClientSideReleaseKey ) , someClientIp , someMessagesAsString , someRequest , someResponse ) ; "<AssertPlaceHolder>" ; verify ( someResponse , times ( 1 ) ) . setStatus ( HttpServletResponse . SC_NOT_MODIFIED ) ; } queryConfig ( java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , javax . servlet . http . HttpServletRequest , javax . servlet . http . HttpServletResponse ) { java . lang . String originalNamespace = namespace ; namespace = namespaceUtil . filterNamespaceName ( namespace ) ; namespace = namespaceUtil . normalizeNamespace ( appId , namespace ) ; if ( com . google . common . base . Strings . isNullOrEmpty ( clientIp ) ) { clientIp = tryToGetClientIp ( request ) ; } com . ctrip . framework . apollo . core . dto . ApolloNotificationMessages clientMessages = transformMessages ( messagesAsString ) ; java . util . List < com . ctrip . framework . apollo . biz . entity . Release > releases = com . google . common . collect . Lists . newLinkedList ( ) ; java . lang . String appClusterNameLoaded = clusterName ; if ( ! ( ConfigConsts . NO_APPID_PLACEHOLDER . equalsIgnoreCase ( appId ) ) ) { com . ctrip . framework . apollo . biz . entity . Release currentAppRelease = configService . loadConfig ( appId , clientIp , appId , clusterName , namespace , dataCenter , clientMessages ) ; if ( currentAppRelease != null ) { releases . add ( currentAppRelease ) ; appClusterNameLoaded = currentAppRelease . getClusterName ( ) ; } } if ( ! ( namespaceBelongsToAppId ( appId , namespace ) ) ) { com . ctrip . framework . apollo . biz . entity . Release publicRelease = this . findPublicConfig ( appId , clientIp , clusterName , namespace , dataCenter , clientMessages ) ; if ( ! ( java . util . Objects . isNull ( publicRelease ) ) ) { releases . add ( publicRelease ) ; } } if ( releases . isEmpty ( ) ) { response . sendError ( HttpServletResponse . SC_NOT_FOUND , java . lang . String . format ( "Could<sp>not<sp>load<sp>configurations<sp>with<sp>appId:<sp>%s,<sp>clusterName:<sp>%s,<sp>namespace:<sp>%s" , appId , clusterName , originalNamespace ) ) ; com . ctrip . framework . apollo . tracer . Tracer . logEvent ( "Apollo.Config.NotFound" , assembleKey ( appId , clusterName , originalNamespace , dataCenter ) ) ; return null ; } auditReleases ( appId , clusterName , dataCenter , clientIp , releases ) ; java . lang . String mergedReleaseKey = releases . stream ( ) . map ( Release :: getReleaseKey ) . collect ( java . util . stream . Collectors . joining ( ConfigConsts . CLUSTER_NAMESPACE_SEPARATOR ) ) ; if ( mergedReleaseKey . equals ( clientSideReleaseKey ) ) { response . setStatus ( HttpServletResponse . SC_NOT_MODIFIED ) ; com . ctrip . framework . apollo . tracer . Tracer . logEvent ( "Apollo.Config.NotModified" , assembleKey ( appId , appClusterNameLoaded , originalNamespace , dataCenter ) ) ; return null ; } com . ctrip . framework . apollo . core . dto . ApolloConfig apolloConfig = new com . ctrip . framework . apollo . core . dto . ApolloConfig ( appId , appClusterNameLoaded , originalNamespace , mergedReleaseKey ) ; apolloConfig . setConfigurations ( mergeReleaseConfigurations ( releases ) ) ; com . ctrip . framework . apollo . tracer . Tracer . logEvent ( "Apollo.Config.Found" , assembleKey ( appId , appClusterNameLoaded , originalNamespace , dataCenter ) ) ; return apolloConfig ; }
org . junit . Assert . assertNull ( result )
centGenerateStepDeltaAbove ( ) { com . google . hashcode . entity . Slice slice = new com . google . hashcode . entity . Slice ( java . util . Arrays . asList ( new com . google . hashcode . entity . Cell ( 0 , 0 , Ingredient . MUSHROOM ) , new com . google . hashcode . entity . Cell ( 0 , 1 , Ingredient . TOMATO ) ) ) ; "<AssertPlaceHolder>" ; } generateStepAbove ( com . google . hashcode . entity . Pizza ) { com . google . hashcode . entity . Slice delta = new com . google . hashcode . entity . Slice ( ) ; for ( int x = this . minX ( ) ; x <= ( this . maxX ( ) ) ; x ++ ) { com . google . hashcode . entity . Optional < com . google . hashcode . entity . Cell > cell = pizza . getCell ( ( ( this . minY ( ) ) - 1 ) , x ) ; if ( cell . isPresent ( ) ) { delta . cells . add ( cell . get ( ) ) ; } else { com . google . hashcode . entity . Slice . LOGGER . debug ( "cant<sp>perform<sp>step<sp>left<sp>!" ) ; return null ; } } com . google . hashcode . entity . Slice . LOGGER . debug ( ( ( "generateStepLeft" + "\nstep<sp>left<sp>delta:<sp>" ) + ( delta . toString ( ) ) ) ) ; com . google . hashcode . entity . Step step = new com . google . hashcode . entity . Step ( this , delta ) ; if ( step . isValid ( pizza ) ) { return step ; } else { com . google . hashcode . entity . Slice . LOGGER . debug ( "step<sp>is<sp>invalid<sp>!" ) ; return null ; } }
org . junit . Assert . assertEquals ( null , slice . generateStepAbove ( pizza ) )
testPopularRepresentativeItemsCSV ( ) { java . lang . String response = target ( "/popularRepresentativeItems" ) . request ( ) . get ( java . lang . String . class ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { java . util . Map < java . lang . String , java . lang . Integer > map = getModel ( ) . getWords ( ) ; java . lang . Integer count ; synchronized ( map ) { count = map . get ( word ) ; } if ( count == null ) { throw new com . cloudera . oryx . api . serving . OryxServingException ( Response . Status . BAD_REQUEST , "No<sp>such<sp>word" ) ; } return count ; }
org . junit . Assert . assertEquals ( 2 , response . split ( "\n" ) . length )
testSimpleStatelessWebserviceEndpoint ( ) { final javax . xml . namespace . QName serviceName = new javax . xml . namespace . QName ( "org.jboss.as.test.integration.ws" , "SimpleService" ) ; final java . net . URL wsdlURL = new java . net . URL ( baseUrl , "/ws-endpoint-example/SimpleService?wsdl" ) ; final javax . xml . ws . Service service = javax . xml . ws . Service . create ( wsdlURL , serviceName ) ; final org . jboss . as . test . integration . ws . SimpleWebserviceEndpointIface port = service . getPort ( org . jboss . as . test . integration . ws . SimpleWebserviceEndpointIface . class ) ; final java . lang . String result = port . echo ( "hello" ) ; "<AssertPlaceHolder>" ; } echo ( java . lang . String ) { org . jboss . as . test . integration . web . sso . interfaces . StatelessSessionBean . log . debug ( ( "echo,<sp>arg=" + arg ) ) ; java . security . Principal p = sessionContext . getCallerPrincipal ( ) ; org . jboss . as . test . integration . web . sso . interfaces . StatelessSessionBean . log . debug ( ( "echo,<sp>callerPrincipal=" + p ) ) ; return p . getName ( ) ; }
org . junit . Assert . assertEquals ( "hello" , result )
testFindPermissionNotifications ( ) { java . lang . String clientId = "APP-5555555555555555" ; java . lang . String orcidId = "some-orcid" ; org . orcid . jaxb . model . notification . permission_v2 . NotificationPermissions notifications = new org . orcid . jaxb . model . notification . permission_v2 . NotificationPermissions ( ) ; notifications . setNotifications ( new java . util . ArrayList ( ) ) ; when ( notificationManager . findPermissionsByOrcidAndClient ( eq ( orcidId ) , eq ( clientId ) , anyInt ( ) , anyInt ( ) ) ) . thenReturn ( notifications ) ; org . orcid . core . utils . SecurityContextTestUtils . setUpSecurityContext ( orcidId , clientId , ScopePathType . PERSON_UPDATE , ScopePathType . PERSON_READ_LIMITED ) ; javax . ws . rs . core . Response response = notificationsApiServiceDelegator . findPermissionNotifications ( "some-orcid" ) ; org . orcid . jaxb . model . notification . permission_v2 . NotificationPermissions retrieved = ( ( org . orcid . jaxb . model . notification . permission_v2 . NotificationPermissions ) ( response . getEntity ( ) ) ) ; "<AssertPlaceHolder>" ; } getNotifications ( ) { return new org . springframework . web . servlet . ModelAndView ( "notifications" ) ; }
org . junit . Assert . assertEquals ( notifications . getNotifications ( ) . size ( ) , retrieved . getNotifications ( ) . size ( ) )
testAndNot ( ) { int [ ] array1 = new int [ ] { 39173 , 39174 , 39175 , 39176 , 39177 , 39178 , 39179 , 39180 , 39181 , 39182 , 39183 , 39184 , 39185 , 39186 , 39187 , 39188 } ; int [ ] array2 = new int [ ] { 14205 } ; org . roaringbitmap . buffer . MutableRoaringBitmap rb1 = org . roaringbitmap . buffer . MutableRoaringBitmap . bitmapOf ( array1 ) ; rb1 . runOptimize ( ) ; org . roaringbitmap . buffer . MutableRoaringBitmap rb2 = org . roaringbitmap . buffer . MutableRoaringBitmap . bitmapOf ( array2 ) ; org . roaringbitmap . buffer . MutableRoaringBitmap answer = org . roaringbitmap . buffer . MutableRoaringBitmap . andNot ( rb1 , rb2 ) ; "<AssertPlaceHolder>" ; } getCardinality ( ) { return ( ( int ) ( getLongCardinality ( ) ) ) ; }
org . junit . Assert . assertEquals ( answer . getCardinality ( ) , array1 . length )
testProxyPutRequest ( ) { org . apache . ambari . server . proxy . ProxyService ps = new org . apache . ambari . server . proxy . ProxyService ( ) ; org . apache . ambari . server . controller . internal . URLStreamProvider streamProviderMock = org . powermock . api . easymock . PowerMock . createNiceMock ( org . apache . ambari . server . controller . internal . URLStreamProvider . class ) ; java . net . HttpURLConnection urlConnectionMock = createMock ( java . net . HttpURLConnection . class ) ; java . net . URI uriMock = org . powermock . api . easymock . PowerMock . createMock ( java . net . URI . class ) ; javax . ws . rs . core . MultivaluedMap < java . lang . String , java . lang . String > queryParams = new com . sun . jersey . core . util . MultivaluedMapImpl ( ) ; javax . ws . rs . core . MultivaluedMap < java . lang . String , java . lang . String > headerParams = new com . sun . jersey . core . util . MultivaluedMapImpl ( ) ; java . util . Map < java . lang . String , java . util . List < java . lang . String > > headerParamsToForward = new java . util . HashMap ( ) ; javax . ws . rs . core . Response . ResponseBuilder responseBuilderMock = org . powermock . api . easymock . PowerMock . createMock ( com . sun . jersey . core . spi . factory . ResponseBuilderImpl . class ) ; javax . ws . rs . core . Response responseMock = createMock ( com . sun . jersey . core . spi . factory . ResponseImpl . class ) ; headerParams . add ( "AmbariProxy-User-Remote" , "testuser" ) ; headerParams . add ( "Content-Type" , "testtype" ) ; java . util . List < java . lang . String > userRemoteParams = new java . util . LinkedList ( ) ; userRemoteParams . add ( "testuser" ) ; headerParamsToForward . put ( "User-Remote" , userRemoteParams ) ; java . io . InputStream is = new java . io . ByteArrayInputStream ( "test" . getBytes ( ) ) ; org . powermock . api . easymock . PowerMock . mockStatic ( javax . ws . rs . core . Response . class ) ; expect ( getHttpHeaders ( ) . getRequestHeaders ( ) ) . andReturn ( headerParams ) ; expect ( getHttpHeaders ( ) . getRequestHeader ( "AmbariProxy-User-Remote" ) ) . andReturn ( userRemoteParams ) ; expect ( getUriInfo ( ) . getRequestUri ( ) ) . andReturn ( uriMock ) ; expect ( getUriInfo ( ) . getQueryParameters ( ) ) . andReturn ( queryParams ) ; expect ( uriMock . getQuery ( ) ) . andReturn ( "url=testurl" ) ; expect ( getHttpHeaders ( ) . getMediaType ( ) ) . andReturn ( javax . ws . rs . core . MediaType . APPLICATION_FORM_URLENCODED_TYPE ) ; expect ( streamProviderMock . processURL ( "testurl" , "PUT" , is , headerParamsToForward ) ) . andReturn ( urlConnectionMock ) ; expect ( urlConnectionMock . getResponseCode ( ) ) . andReturn ( 200 ) ; expect ( urlConnectionMock . getContentType ( ) ) . andReturn ( "text/plain" ) ; expect ( urlConnectionMock . getInputStream ( ) ) . andReturn ( is ) ; org . powermock . api . easymock . PowerMock . expectNew ( org . apache . ambari . server . controller . internal . URLStreamProvider . class , 20000 , 15000 , null , null , null ) . andReturn ( streamProviderMock ) ; expect ( javax . ws . rs . core . Response . status ( 200 ) ) . andReturn ( responseBuilderMock ) ; expect ( responseBuilderMock . entity ( is ) ) . andReturn ( responseBuilderMock ) ; expect ( responseBuilderMock . type ( "text/plain" ) ) . andReturn ( responseBuilderMock ) ; expect ( responseBuilderMock . build ( ) ) . andReturn ( responseMock ) ; org . powermock . api . easymock . PowerMock . replay ( streamProviderMock , org . apache . ambari . server . controller . internal . URLStreamProvider . class , javax . ws . rs . core . Response . class , responseBuilderMock , uriMock , java . net . URI . class ) ; replay ( getUriInfo ( ) , urlConnectionMock , getHttpHeaders ( ) ) ; javax . ws . rs . core . Response resultForPutRequest = ps . processPutRequestForwarding ( is , getHttpHeaders ( ) , getUriInfo ( ) ) ; "<AssertPlaceHolder>" ; } getUriInfo ( ) { return uriInfo ; }
org . junit . Assert . assertSame ( resultForPutRequest , responseMock )
testNullObjectHash ( ) { final int h1 = org . apache . hc . core5 . util . LangUtils . hashCode ( LangUtils . HASH_SEED , null ) ; final int h2 = org . apache . hc . core5 . util . LangUtils . hashCode ( LangUtils . HASH_SEED , 0 ) ; "<AssertPlaceHolder>" ; } hashCode ( int , int ) { return ( seed * ( org . apache . hc . core5 . util . LangUtils . HASH_OFFSET ) ) + hashcode ; }
org . junit . Assert . assertTrue ( ( h1 == h2 ) )
testJUnitHamcrestMatcherFailureWorks ( ) { try { "<AssertPlaceHolder>" ; } catch ( java . lang . NoSuchMethodError e ) { org . junit . Assert . fail ( ( ( "Class<sp>search<sp>path<sp>seems<sp>broken<sp>re<sp>new<sp>JUnit<sp>and<sp>old<sp>Hamcrest." + "<sp>Got<sp>NoSuchMethodError;<sp>e:<sp>" ) + e ) ) ; } catch ( java . lang . AssertionError e ) { System . out . println ( ( "Class<sp>path<sp>seems<sp>fine<sp>re<sp>new<sp>JUnit<sp>vs.<sp>old<sp>Hamcrest." + "<sp>(Got<sp>AssertionError,<sp>not<sp>NoSuchMethodError.)" ) ) ; } }
org . junit . Assert . assertThat ( 1 , org . hamcrest . CoreMatchers . equalTo ( 2 ) )
testGetRowKeys ( ) { org . jfree . data . category . DefaultIntervalCategoryDataset empty = new org . jfree . data . category . DefaultIntervalCategoryDataset ( new double [ 0 ] [ 0 ] , new double [ 0 ] [ 0 ] ) ; java . util . List keys = empty . getRowKeys ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . tickUnits . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , keys . size ( ) )
testGetNextTermBegin ( ) { java . util . Date startDate = java . util . Date . from ( java . time . ZonedDateTime . ofLocal ( java . time . LocalDate . of ( 2016 , 9 , 23 ) . atStartOfDay ( ) , java . time . ZoneId . systemDefault ( ) , null ) . toInstant ( ) ) ; de . janrieke . contractmanager . util . ValidRuntimes runtimes = new de . janrieke . contractmanager . util . ValidRuntimes ( ) ; runtimes . firstMinRuntimeCount = 6 ; runtimes . firstMinRuntimeType = de . janrieke . contractmanager . rmi . IntervalType . MONTHS ; runtimes . followingMinRuntimeCount = 3 ; runtimes . followingMinRuntimeType = de . janrieke . contractmanager . rmi . IntervalType . MONTHS ; java . util . Date actual = de . janrieke . contractmanager . util . DateUtils . calculateNextTermBeginAfter ( now , startDate , false , runtimes ) ; java . util . Date expected = java . util . Date . from ( java . time . ZonedDateTime . ofLocal ( java . time . LocalDate . of ( 2017 , 3 , 23 ) . atStartOfDay ( ) , java . time . ZoneId . systemDefault ( ) , null ) . toInstant ( ) ) ; "<AssertPlaceHolder>" ; } calculateNextTermBeginAfter ( java . util . Date , java . util . Date , boolean , de . janrieke . contractmanager . util . ValidRuntimes ) { if ( after == null ) { return null ; } if ( startDate == null ) { return null ; } if ( runtimes == null ) { return null ; } java . util . Calendar afterCal = java . util . Calendar . getInstance ( ) ; afterCal . setTime ( de . willuhn . jameica . util . DateUtil . endOfDay ( after ) ) ; java . util . Calendar calendar = java . util . Calendar . getInstance ( ) ; calendar . setTime ( startDate ) ; if ( runtimes . fixedTerms ) { switch ( runtimes . firstMinRuntimeType ) { case WEEKS : calendar . set ( Calendar . DAY_OF_WEEK , Calendar . MONDAY ) ; calendar . add ( Calendar . WEEK_OF_YEAR , 1 ) ; break ; case MONTHS : calendar . set ( Calendar . DAY_OF_MONTH , 1 ) ; calendar . add ( Calendar . MONTH , 1 ) ; break ; case YEARS : calendar . set ( Calendar . DAY_OF_YEAR , 1 ) ; calendar . add ( Calendar . YEAR , 1 ) ; break ; default : } } boolean first = true ; boolean validResult = true ; while ( validResult && ( ( ! ( calendar . after ( afterCal ) ) ) || ( first && excludeFirstTerm ) ) ) { if ( first ) { validResult = de . janrieke . contractmanager . util . DateUtils . addToCalendar ( calendar , runtimes . firstMinRuntimeType , runtimes . firstMinRuntimeCount ) ; first = false ; } else { validResult = de . janrieke . contractmanager . util . DateUtils . addToCalendar ( calendar , runtimes . followingMinRuntimeType , runtimes . followingMinRuntimeCount ) ; } } return validResult ? calendar . getTime ( ) : null ; }
org . junit . Assert . assertEquals ( expected , actual )
withParserFactory ( ) { com . github . jknack . handlebars . ParserFactory parserFactory = createMock ( com . github . jknack . handlebars . ParserFactory . class ) ; replay ( parserFactory ) ; com . github . jknack . handlebars . Handlebars handlebars = new com . github . jknack . handlebars . Handlebars ( ) . with ( parserFactory ) ; "<AssertPlaceHolder>" ; verify ( parserFactory ) ; } with ( com . github . jknack . handlebars . io . TemplateLoader [ ] ) { org . apache . commons . lang3 . Validate . isTrue ( ( ( loader . length ) > 0 ) , "The<sp>template<sp>loader<sp>is<sp>required." ) ; this . loader = ( ( loader . length ) == 1 ) ? loader [ 0 ] : new com . github . jknack . handlebars . io . CompositeTemplateLoader ( loader ) ; return this ; }
org . junit . Assert . assertNotNull ( handlebars )
testGetDisplayEvents ( ) { org . joda . time . DateTimeZone tz = org . joda . time . DateTimeZone . forID ( "America/Los_Angeles" ) ; org . joda . time . DateTime eventStart = new org . joda . time . DateTime ( 2012 , 1 , 4 , 17 , 0 , tz ) ; org . joda . time . DateTime eventEnd = new org . joda . time . DateTime ( 2012 , 1 , 4 , 18 , 0 , tz ) ; net . fortuna . ical4j . model . component . VEvent event = new net . fortuna . ical4j . model . component . VEvent ( getICal4jDate ( eventStart , tz ) , getICal4jDate ( eventEnd , tz ) , "Test<sp>Event" ) ; org . joda . time . DateMidnight intervalStart = new org . joda . time . DateMidnight ( 2012 , 1 , 3 , tz ) ; org . joda . time . DateMidnight intervalStop = new org . joda . time . DateMidnight ( 2012 , 1 , 5 , tz ) ; org . joda . time . Interval interval = new org . joda . time . Interval ( intervalStart , intervalStop ) ; java . util . Set < org . jasig . portlet . calendar . mvc . CalendarDisplayEvent > events = eventDao . getDisplayEvents ( event , interval , Locale . US , tz ) ; "<AssertPlaceHolder>" ; } getDisplayEvents ( net . fortuna . ical4j . model . component . VEvent , org . joda . time . Interval , java . util . Locale , org . joda . time . DateTimeZone ) { final net . fortuna . ical4j . model . component . VEvent event = ( ( net . fortuna . ical4j . model . component . VEvent ) ( e . copy ( ) ) ) ; org . joda . time . DateTime eventStart ; org . joda . time . DateTime eventEnd = null ; if ( ( ( event . getStartDate ( ) . getTimeZone ( ) ) == null ) && ( ! ( event . getStartDate ( ) . isUtc ( ) ) ) ) { if ( log . isDebugEnabled ( ) ) { log . debug ( ( ( "Identified<sp>event<sp>" + ( event . getSummary ( ) ) ) + "<sp>as<sp>a<sp>floating<sp>event" ) ) ; } int offset = usersConfiguredDateTimeZone . getOffset ( event . getStartDate ( ) . getDate ( ) . getTime ( ) ) ; eventStart = new org . joda . time . DateTime ( ( ( event . getStartDate ( ) . getDate ( ) . getTime ( ) ) - offset ) , usersConfiguredDateTimeZone ) ; if ( ( event . getEndDate ( ) ) != null ) { eventEnd = new org . joda . time . DateTime ( ( ( event . getEndDate ( ) . getDate ( ) . getTime ( ) ) - offset ) , usersConfiguredDateTimeZone ) ; } } else { eventStart = new org . joda . time . DateTime ( event . getStartDate ( ) . getDate ( ) , usersConfiguredDateTimeZone ) ; if ( ( event . getEndDate ( ) ) != null ) { eventEnd = new org . joda . time . DateTime ( event . getEndDate ( ) . getDate ( ) , usersConfiguredDateTimeZone ) ; } } if ( eventEnd == null ) { eventEnd = eventStart ; } final java . util . Date firstDayToProcess = ( interval . contains ( event . getStartDate ( ) . getDate ( ) . getTime ( ) ) ) ? event . getStartDate ( ) . getDate ( ) : interval . getStart ( ) . toDate ( ) ; org . joda . time . DateMidnight startOfTheSpecificDay = new org . joda . time . DateMidnight ( firstDayToProcess , usersConfiguredDateTimeZone ) ; org . joda . time . DateMidnight endOfTheSpecificDay = startOfTheSpecificDay . plusDays ( 1 ) ; final org . joda . time . format . DateTimeFormatter df = getDateFormatter ( locale , usersConfiguredDateTimeZone ) ; final org . joda . time . format . DateTimeFormatter tf = getTimeFormatter ( locale , usersConfiguredDateTimeZone ) ; final java . util . Set < org . jasig . portlet . calendar . mvc . CalendarDisplayEvent > events = new java . util . HashSet < org . jasig . portlet . calendar . mvc . CalendarDisplayEvent > ( ) ; final org . joda . time . Interval eventInterval = new org . joda . time . Interval ( eventStart , eventEnd ) ; do { final org . joda . time . Interval theSpecificDay = new org . joda . time . Interval ( startOfTheSpecificDay . getMillis ( ) , endOfTheSpecificDay . getMillis ( ) , usersConfiguredDateTimeZone ) ; if ( ( theSpecificDay . getStart ( ) . isEqual ( eventStart ) ) || ( theSpecificDay . overlaps ( eventInterval ) ) ) { final org . jasig . portlet . calendar . mvc . CalendarDisplayEvent json = new org . jasig . portlet . calendar . mvc . CalendarDisplayEvent ( event , eventInterval , theSpecificDay , df , tf ) ; events . add ( json ) ; } startOfTheSpecificDay = startOfTheSpecificDay . plusDays ( 1 ) ; endOfTheSpecificDay = endOfTheSpecificDay . plusDays ( 1 ) ; } while ( ( ! ( startOfTheSpecificDay . isAfter ( eventEnd ) ) ) && ( interval . contains ( startOfTheSpecificDay ) ) ) ; return events ; }
org . junit . Assert . assertEquals ( 1 , events . size ( ) )
testGetSessionProperties ( ) { final org . apache . flink . table . client . gateway . Executor executor = createDefaultExecutor ( org . apache . flink . table . client . gateway . local . LocalExecutorITCase . clusterClient ) ; final org . apache . flink . table . client . gateway . SessionContext session = new org . apache . flink . table . client . gateway . SessionContext ( "execution.max-table-result-rows" 8 , new org . apache . flink . table . client . config . Environment ( ) ) ; session . setSessionProperty ( "99000" 5 , "execution.max-table-result-rows" 7 ) ; executor . getSessionProperties ( session ) ; session . setSessionProperty ( "99000" 5 , "99000" 4 ) ; final java . util . Map < java . lang . String , java . lang . String > actualProperties = executor . getSessionProperties ( session ) ; final java . util . Map < java . lang . String , java . lang . String > expectedProperties = new java . util . HashMap ( ) ; expectedProperties . put ( "99000" 8 , "99000" 1 ) ; expectedProperties . put ( "99000" 7 , "execution.max-table-result-rows" 0 ) ; expectedProperties . put ( "execution.periodic-watermarks-interval" , "99000" 6 ) ; expectedProperties . put ( "execution.max-table-result-rows" 2 , "99000" 9 ) ; expectedProperties . put ( "execution.max-parallelism" , "16" ) ; expectedProperties . put ( "execution.max-table-result-rows" 6 , "99000" 2 ) ; expectedProperties . put ( "execution.max-table-result-rows" 3 , "99000" 2 ) ; expectedProperties . put ( "99000" 5 , "99000" 4 ) ; expectedProperties . put ( "execution.max-table-result-rows" , "99000" 3 ) ; expectedProperties . put ( "execution.restart-strategy.type" , "failure-rate" ) ; expectedProperties . put ( "execution.max-table-result-rows" 5 , "execution.max-table-result-rows" 1 ) ; expectedProperties . put ( "99000" 0 , "99000" ) ; expectedProperties . put ( "execution.restart-strategy.delay" , "1000" ) ; expectedProperties . put ( "execution.max-table-result-rows" 4 , "5000" ) ; "<AssertPlaceHolder>" ; } put ( org . apache . flink . api . common . JobID , org . apache . flink . runtime . taskexecutor . JobManagerConnection ) { org . apache . flink . runtime . taskexecutor . JobManagerConnection previousJMC = jobManagerConnections . put ( jobId , jobManagerConnection ) ; if ( previousJMC != null ) { jobManagerConnections . put ( jobId , previousJMC ) ; return false ; } else { return true ; } }
org . junit . Assert . assertEquals ( expectedProperties , actualProperties )
shouldAlwaysLocateTheElementPerCall ( ) { final org . openqa . selenium . support . pagefactory . ElementLocator locator = mock ( org . openqa . selenium . support . pagefactory . ElementLocator . class ) ; final org . openqa . selenium . WebElement element1 = mock ( org . openqa . selenium . WebElement . class , "webElement1" ) ; final org . openqa . selenium . WebElement element2 = mock ( org . openqa . selenium . WebElement . class , "webElement2" ) ; final java . util . List < org . openqa . selenium . WebElement > list = java . util . Arrays . asList ( element1 , element2 ) ; when ( locator . findElements ( ) ) . thenReturn ( list ) ; org . openqa . selenium . support . pagefactory . internal . LocatingElementListHandler handler = new org . openqa . selenium . support . pagefactory . internal . LocatingElementListHandler ( locator ) ; java . util . List < org . openqa . selenium . WebElement > proxy = ( ( java . util . List < org . openqa . selenium . WebElement > ) ( java . lang . reflect . Proxy . newProxyInstance ( getClass ( ) . getClassLoader ( ) , new java . lang . Class [ ] { java . util . List . class } , handler ) ) ) ; proxy . get ( 1 ) . sendKeys ( "Fishy" ) ; "<AssertPlaceHolder>" ; verify ( locator , times ( 2 ) ) . findElements ( ) ; verify ( element2 , times ( 1 ) ) . sendKeys ( "Fishy" ) ; verifyNoMoreInteractions ( locator , element2 ) ; verifyZeroInteractions ( element1 ) ; } size ( ) { java . lang . Object response = executeMethod . execute ( DriverCommand . GET_LOCAL_STORAGE_SIZE , null ) ; return java . lang . Integer . parseInt ( response . toString ( ) ) ; }
org . junit . Assert . assertThat ( proxy . size ( ) , org . hamcrest . Matchers . equalTo ( 2 ) )
testReadWrite ( ) { com . bah . culvert . constraints . IndexRangeConstraint ct = new com . bah . culvert . constraints . IndexRangeConstraint ( new com . bah . culvert . mock . MockIndex ( ) , new com . bah . culvert . data . CRange ( new byte [ ] { 2 } ) ) ; ct = ( ( com . bah . culvert . constraints . IndexRangeConstraint ) ( com . bah . culvert . test . Utils . testReadWrite ( ct ) ) ) ; "<AssertPlaceHolder>" ; } getRange ( ) { return this . rowRange ; }
org . junit . Assert . assertEquals ( new com . bah . culvert . data . CRange ( new byte [ ] { 2 } ) , ct . getRange ( ) )
givenListWithElements_whenToArray_thenArrayContainsThose ( ) { java . util . List < java . lang . Object > list = new com . baeldung . java . list . CustomList ( ) ; list . add ( "baeldung" ) ; list . add ( ".com" ) ; java . lang . Object [ ] array = list . toArray ( ) ; "<AssertPlaceHolder>" ; } toArray ( ) { return java . util . Arrays . copyOf ( internal , internal . length ) ; }
org . junit . Assert . assertArrayEquals ( new java . lang . Object [ ] { "baeldung" , ".com" } , array )
testDeleteIdsThrow ( ) { com . j256 . ormlite . dao . Dao < com . j256 . ormlite . dao . Foo , java . lang . Integer > dao = createDao ( com . j256 . ormlite . dao . Foo . class , true ) ; com . j256 . ormlite . dao . Foo foo = new com . j256 . ormlite . dao . Foo ( ) ; "<AssertPlaceHolder>" ; com . j256 . ormlite . support . DatabaseConnection conn = connectionSource . getReadWriteConnection ( com . j256 . ormlite . dao . FOO_TABLE_NAME ) ; try { conn . close ( ) ; java . util . List < java . lang . Integer > foos = new java . util . ArrayList < java . lang . Integer > ( ) ; foos . add ( foo . id ) ; dao . deleteIds ( foos ) ; } finally { connectionSource . releaseConnection ( conn ) ; } } create ( T ) { checkForInitialized ( ) ; if ( data == null ) { return 0 ; } if ( data instanceof com . j256 . ormlite . misc . BaseDaoEnabled ) { @ com . j256 . ormlite . dao . SuppressWarnings ( "unchecked" ) com . j256 . ormlite . misc . BaseDaoEnabled < T , ID > daoEnabled = ( ( com . j256 . ormlite . misc . BaseDaoEnabled < T , ID > ) ( data ) ) ; daoEnabled . setDao ( this ) ; } com . j256 . ormlite . support . DatabaseConnection connection = connectionSource . getReadWriteConnection ( tableInfo . getTableName ( ) ) ; try { return statementExecutor . create ( connection , data , objectCache ) ; } finally { connectionSource . releaseConnection ( connection ) ; } }
org . junit . Assert . assertEquals ( 1 , dao . create ( foo ) )
testMapToString_error ( ) { com . cybersource . ws . client . Map map = null ; java . lang . String result = com . cybersource . ws . client . Utility . mapToString ( map , false , 0 ) ; "<AssertPlaceHolder>" ; } mapToString ( java . util . Map , boolean , int ) { java . lang . StringBuffer dest = new java . lang . StringBuffer ( ) ; if ( ( src != null ) && ( ! ( src . isEmpty ( ) ) ) ) { java . util . Iterator iter = src . keySet ( ) . iterator ( ) ; java . lang . String key ; java . lang . String val ; while ( iter . hasNext ( ) ) { key = ( ( java . lang . String ) ( iter . next ( ) ) ) ; val = ( mask ) ? com . cybersource . ws . client . PCI . maskIfNotSafe ( type , key , ( ( java . lang . String ) ( src . get ( key ) ) ) ) : ( ( java . lang . String ) ( src . get ( key ) ) ) ; dest . append ( ( ( ( key + "=" ) + val ) + "\n" ) ) ; } } else { return dest . toString ( ) ; } java . lang . String hasEscapes = ( ( java . lang . String ) ( src . get ( com . cybersource . ws . client . Utility . HAS_ESCAPES ) ) ) ; return ( "1" . equals ( hasEscapes ) ) || ( "true" . equalsIgnoreCase ( hasEscapes ) ) ? dest . toString ( ) : org . apache . commons . lang3 . StringEscapeUtils . escapeXml11 ( dest . toString ( ) ) ; }
org . junit . Assert . assertTrue ( result . isEmpty ( ) )
testGetPreviousVisibleColumn_NullArgument ( ) { org . eclipse . nebula . widgets . grid . GridColumn [ ] columns = org . eclipse . nebula . widgets . grid . GridTestUtil . createGridColumns ( grid , 5 , SWT . NONE ) ; "<AssertPlaceHolder>" ; } getPreviousVisibleColumn ( org . eclipse . nebula . widgets . grid . GridColumn ) { checkWidget ( ) ; org . eclipse . nebula . widgets . grid . GridColumn result = null ; int index = 0 ; if ( column == null ) { index = displayOrderedColumns . size ( ) ; } else { index = displayOrderedColumns . indexOf ( column ) ; } if ( index > 0 ) { result = displayOrderedColumns . get ( ( index - 1 ) ) ; while ( ( result != null ) && ( ! ( result . isVisible ( ) ) ) ) { index -- ; if ( index > 0 ) { result = displayOrderedColumns . get ( ( index - 1 ) ) ; } else { result = null ; } } } return result ; }
org . junit . Assert . assertSame ( columns [ 4 ] , grid . getPreviousVisibleColumn ( null ) )
unmodifiable2 ( ) { final com . github . rinde . rinsim . geom . Point N = new com . github . rinde . rinsim . geom . Point ( 0 , 5 ) ; final com . github . rinde . rinsim . geom . Point E = new com . github . rinde . rinsim . geom . Point ( 5 , 0 ) ; final com . github . rinde . rinsim . geom . Point S = new com . github . rinde . rinsim . geom . Point ( 0 , ( - 5 ) ) ; final com . github . rinde . rinsim . geom . Point W = new com . github . rinde . rinsim . geom . Point ( ( - 5 ) , 0 ) ; com . github . rinde . rinsim . geom . Graphs . addBiPath ( graph , N , E , S , W , N ) ; final com . github . rinde . rinsim . geom . Graph < com . github . rinde . rinsim . geom . LengthData > unmod = com . github . rinde . rinsim . geom . Graphs . unmodifiableGraph ( graph ) ; graph . addConnection ( N , S ) ; "<AssertPlaceHolder>" ; } getConnection ( com . github . rinde . rinsim . geom . Point , com . github . rinde . rinsim . geom . Point ) { checkArgument ( hasConnection ( from , to ) , "%s<sp>-><sp>%s<sp>is<sp>not<sp>a<sp>connection" , from , to ) ; return data . get ( from , to ) ; }
org . junit . Assert . assertEquals ( graph . getConnection ( N , S ) , unmod . getConnection ( N , S ) )
telecomNullTest ( ) { org . marc . everest . datatypes . generic . SET < org . marc . everest . datatypes . TEL > telecoms = org . oscarehr . e2e . model . export . header . RecordTargetModelTest . nullRecordTargetModel . getTelecoms ( ) ; "<AssertPlaceHolder>" ; } getTelecoms ( ) { return telecoms ; }
org . junit . Assert . assertNull ( telecoms )
testGetIsLongOrDouble ( ) { org . pentaho . hbase . shim . api . HBaseValueMeta hbMeta = getHBaseValueMeta ( ) ; hbMeta . m_isLongOrDouble = true ; "<AssertPlaceHolder>" ; } getIsLongOrDouble ( ) { return m_isLongOrDouble ; }
org . junit . Assert . assertTrue ( hbMeta . getIsLongOrDouble ( ) )
nextLevel_skip_currentMoreThanTotal ( ) { json = new org . json . JSONObject ( "{'levelProgress':{'total':2,'current':2,'lastPassed':2}}" ) ; json = com . codenjoy . dojo . services . multiplayer . LevelProgress . winLevel ( json ) ; "<AssertPlaceHolder>" ; } winLevel ( org . json . JSONObject ) { com . codenjoy . dojo . services . multiplayer . LevelProgress progress = new com . codenjoy . dojo . services . multiplayer . LevelProgress ( json ) ; org . json . JSONObject clone = new org . json . JSONObject ( json . toString ( ) ) ; progress . change ( ( ( progress . current ) + 1 ) , java . lang . Math . max ( progress . passed , progress . current ) ) ; if ( progress . isValid ( ) ) { return progress . saveTo ( clone ) ; } else { return null ; } }
org . junit . Assert . assertEquals ( null , json )
convertToCnfTest ( ) { org . s1ck . gdl . model . predicates . expressions . Comparison a = getComparison ( ) ; org . s1ck . gdl . model . predicates . expressions . Comparison b = getComparison ( ) ; org . gradoop . flink . model . impl . operators . matching . common . query . predicates . booleans . AndPredicate andPredicate = new org . gradoop . flink . model . impl . operators . matching . common . query . predicates . booleans . AndPredicate ( new org . s1ck . gdl . model . predicates . booleans . And ( a , b ) ) ; org . gradoop . flink . model . impl . operators . matching . common . query . predicates . CNF reference = andPredicate . getLhs ( ) . asCNF ( ) . and ( andPredicate . getRhs ( ) . asCNF ( ) ) ; "<AssertPlaceHolder>" ; } asCNF ( ) { org . gradoop . flink . model . impl . operators . matching . common . query . predicates . CNF cnf = new org . gradoop . flink . model . impl . operators . matching . common . query . predicates . CNF ( ) ; org . gradoop . flink . model . impl . operators . matching . common . query . predicates . CNFElement cnfElement = new org . gradoop . flink . model . impl . operators . matching . common . query . predicates . CNFElement ( ) ; cnfElement . addPredicate ( this ) ; cnf . addPredicate ( cnfElement ) ; return cnf ; }
org . junit . Assert . assertEquals ( reference , andPredicate . asCNF ( ) )
shouldReturnFalseIfOtherVersionIsNull ( ) { com . github . zafarkhaja . semver . MetadataVersion v1 = new com . github . zafarkhaja . semver . MetadataVersion ( new java . lang . String [ ] { "alpha" , "123" } ) ; com . github . zafarkhaja . semver . MetadataVersion v2 = null ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == other ) { return true ; } if ( ! ( other instanceof com . github . zafarkhaja . semver . Version ) ) { return false ; } return ( compareTo ( ( ( com . github . zafarkhaja . semver . Version ) ( other ) ) ) ) == 0 ; }
org . junit . Assert . assertFalse ( v1 . equals ( v2 ) )
testGetNumReducersBadPath ( ) { com . nearinfinity . blur . mapreduce . BlurTask task = new com . nearinfinity . blur . mapreduce . BlurTask ( ) ; com . nearinfinity . blur . thrift . generated . TableDescriptor tableDescriptor = new com . nearinfinity . blur . thrift . generated . TableDescriptor ( ) ; tableDescriptor . setShardCount ( 5 ) ; tableDescriptor . setTableUri ( "file:///tmp/blur34746545" ) ; tableDescriptor . setName ( "blur34746545" ) ; task . setTableDescriptor ( tableDescriptor ) ; "<AssertPlaceHolder>" ; } getNumReducers ( org . apache . hadoop . conf . Configuration ) { org . apache . hadoop . fs . Path tablePath = new org . apache . hadoop . fs . Path ( _tableDescriptor . tableUri ) ; try { int num = _tableDescriptor . shardCount ; org . apache . hadoop . fs . FileSystem fileSystem = org . apache . hadoop . fs . FileSystem . get ( tablePath . toUri ( ) , configuration ) ; if ( ! ( fileSystem . exists ( tablePath ) ) ) { return num ; } org . apache . hadoop . fs . FileStatus [ ] files = fileSystem . listStatus ( tablePath ) ; int shardCount = 0 ; for ( org . apache . hadoop . fs . FileStatus fileStatus : files ) { if ( fileStatus . isDir ( ) ) { java . lang . String name = fileStatus . getPath ( ) . getName ( ) ; if ( name . startsWith ( BlurConstants . SHARD_PREFIX ) ) { shardCount ++ ; } } } if ( shardCount == 0 ) { return num ; } if ( shardCount != num ) { com . nearinfinity . blur . mapreduce . BlurTask . LOG . warn ( ( ( ( ( ( ( ( ( "Asked<sp>for<sp>" + num ) + "<sp>reducers,<sp>but<sp>existing<sp>table<sp>" ) + ( _tableDescriptor . name ) ) + "<sp>has<sp>" ) + shardCount ) + "<sp>shards.<sp>Using<sp>" ) + shardCount ) + "<sp>reducers" ) ) ; } return shardCount ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( "Unable<sp>to<sp>connect<sp>to<sp>filesystem" , e ) ; } }
org . junit . Assert . assertEquals ( 5 , task . getNumReducers ( new org . apache . hadoop . conf . Configuration ( ) ) )
testBerichtSyntaxException ( ) { final java . lang . String berichtOrigineel = org . apache . commons . io . IOUtils . toString ( nl . moderniseringgba . isc . esb . message . brp . impl . MvGeboorteVerzoekBerichtTest . class . getResourceAsStream ( "mvGeboorteVerzoekBerichtSyntaxExceptionBericht.xml" ) ) ; final nl . moderniseringgba . isc . esb . message . brp . BrpBericht brpBericht = factory . getBericht ( berichtOrigineel ) ; "<AssertPlaceHolder>" ; } getBericht ( java . lang . String ) { try { final javax . xml . bind . JAXBElement < ? > element = NotificatieXml . SINGLETON . stringToElement ( berichtAlsString ) ; return maakBericht ( element . getValue ( ) ) ; } catch ( final javax . xml . bind . JAXBException e ) { nl . bzk . migratiebrp . bericht . model . notificatie . factory . NotificatieBerichtFactory . LOG . warn ( "Verwerken<sp>bericht<sp>mislukt" , e ) ; return new nl . bzk . migratiebrp . bericht . model . notificatie . impl . OngeldigBericht ( berichtAlsString , e . getMessage ( ) ) ; } }
org . junit . Assert . assertTrue ( ( brpBericht instanceof nl . moderniseringgba . isc . esb . message . brp . impl . OngeldigBericht ) )
whenFileIsNotModified_getFileModeShouldReturnItsOriginalFileMode ( ) { writeToCache ( "/some_file.txt" , someBytes ( ) , com . beijunyi . parallelgit . filesystem . io . EXECUTABLE_FILE ) ; initGitFileSystem ( ) ; "<AssertPlaceHolder>" ; } fileMode ( java . lang . String ) { return ( ( org . eclipse . jgit . lib . FileMode ) ( readAttribute ( path , com . beijunyi . parallelgit . filesystem . io . FILE_MODE ) ) ) ; }
org . junit . Assert . assertEquals ( com . beijunyi . parallelgit . filesystem . io . EXECUTABLE_FILE , fileMode ( "/some_file.txt" ) )
testEquals ( ) { org . jfree . chart . plot . CombinedDomainCategoryPlot plot1 = createPlot ( ) ; org . jfree . chart . plot . CombinedDomainCategoryPlot plot2 = createPlot ( ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( obj == ( this ) ) { return true ; } if ( ! ( obj instanceof org . jfree . data . xy . YWithXInterval ) ) { return false ; } org . jfree . data . xy . YWithXInterval that = ( ( org . jfree . data . xy . YWithXInterval ) ( obj ) ) ; if ( ( this . y ) != ( that . y ) ) { return false ; } if ( ( this . xLow ) != ( that . xLow ) ) { return false ; } if ( ( this . xHigh ) != ( that . xHigh ) ) { return false ; } return true ; }
org . junit . Assert . assertTrue ( plot1 . equals ( plot2 ) )
testIsStableVersion_null ( ) { "<AssertPlaceHolder>" ; } isStableVersion ( net . sourceforge . jwbf . mediawiki . MediaWiki$Version ) { if ( version != null ) { java . lang . reflect . Field field = net . sourceforge . jwbf . mediawiki . MediaWiki . Version . getField ( version ) ; boolean isDeprecated = field . isAnnotationPresent ( net . sourceforge . jwbf . mediawiki . Deprecated . class ) ; boolean isBeta = field . isAnnotationPresent ( net . sourceforge . jwbf . core . NotReleased . class ) ; return ! ( ( ( ( version . equals ( net . sourceforge . jwbf . mediawiki . MediaWiki . Version . DEVELOPMENT ) ) || ( version . equals ( net . sourceforge . jwbf . mediawiki . MediaWiki . Version . UNKNOWN ) ) ) || isDeprecated ) || isBeta ) ; } else { return false ; } }
org . junit . Assert . assertFalse ( MediaWiki . Version . isStableVersion ( null ) )
testSuccess ( ) { com . alibaba . craftsman . dto . CodeReviewMetricAddCmd codeReviewMetricAddCmd = com . alibaba . craftsman . app . CodeReviewMetricAddCmdExeTest . prepareCodeReviewMetricAddCmd ( "CodeReviewMetricAddCmdExeTest_098873" ) ; com . alibaba . cola . dto . Response response = metricsService . addCodeReviewMetric ( codeReviewMetricAddCmd ) ; "<AssertPlaceHolder>" ; } isSuccess ( ) { return isSuccess ; }
org . junit . Assert . assertTrue ( response . isSuccess ( ) )
testGetTotalCurrentMips ( ) { java . util . List < java . lang . Double > mipsShare = new java . util . ArrayList < java . lang . Double > ( ) ; mipsShare . add ( ( ( org . cloudbus . cloudsim . CloudletSchedulerSingleServiceTest . MIPS ) / 4 ) ) ; mipsShare . add ( ( ( org . cloudbus . cloudsim . CloudletSchedulerSingleServiceTest . MIPS ) / 4 ) ) ; vmScheduler . setCurrentMipsShare ( mipsShare ) ; "<AssertPlaceHolder>" ; } getTotalCurrentMips ( ) { int totalCurrentMips = 0 ; for ( double mips : getCurrentMipsShare ( ) ) { totalCurrentMips += mips ; } return totalCurrentMips ; }
org . junit . Assert . assertEquals ( ( ( org . cloudbus . cloudsim . CloudletSchedulerSingleServiceTest . MIPS ) / 2 ) , vmScheduler . getTotalCurrentMips ( ) , 0 )
testAddSessionListener1 ( ) { manager . addSessionListener ( listener ) ; com . mountainminds . eclemma . core . ICoverageSession s0 = new com . mountainminds . eclemma . internal . core . SessionManagerTest . DummySession ( ) ; manager . addSession ( s0 , false , null ) ; reflistener . sessionAdded ( s0 ) ; "<AssertPlaceHolder>" ; } sessionAdded ( com . mountainminds . eclemma . core . ICoverageSession ) { }
org . junit . Assert . assertEquals ( reflistener , listener )
testGetTarget ( ) { System . out . println ( "getTarget" ) ; gov . hhs . fha . nhinc . gateway . executorservice . CallableRequest < gov . hhs . fha . nhinc . connectmgr . UrlInfo , org . hl7 . v3 . RespondingGatewayPRPAIN201305UV02RequestType , gov . hhs . fha . nhinc . gateway . executorservice . ResponseWrapper > instance = new gov . hhs . fha . nhinc . gateway . executorservice . CallableRequest ( null , null , null , null ) ; java . lang . Object expResult = null ; java . lang . Object result = instance . getTarget ( ) ; "<AssertPlaceHolder>" ; } getTarget ( ) { return target ; }
org . junit . Assert . assertEquals ( expResult , result )
testDeepCopy_ByteBuffer ( ) { byte [ ] bytes = new byte [ ] { 1 , 2 , 3 } ; java . nio . ByteBuffer buffer = java . nio . ByteBuffer . wrap ( bytes ) ; java . nio . ByteBuffer deepCopied = new org . apache . crunch . types . avro . AvroDeepCopier . AvroByteBufferDeepCopier ( ) . INSTANCE . deepCopy ( buffer ) ; bytes [ 0 ] = 0 ; "<AssertPlaceHolder>" ; } deepCopy ( T ) { return source ; }
org . junit . Assert . assertArrayEquals ( new byte [ ] { 1 , 2 , 3 } , deepCopied . array ( ) )
emptyContext ( ) { org . apache . cayenne . access . translator . select . GroupByStage stage = new org . apache . cayenne . access . translator . select . GroupByStage ( ) ; stage . perform ( context ) ; org . apache . cayenne . access . sqlbuilder . sqltree . Node node = context . getSelectBuilder ( ) . build ( ) ; "<AssertPlaceHolder>" ; } getChildrenCount ( ) { return ( subentities ) != null ? subentities . size ( ) : 0 ; }
org . junit . Assert . assertEquals ( 0 , node . getChildrenCount ( ) )
generateFromAvroEventWithEmptyLogEventsTest ( ) { eventBuilder = new org . kaaproject . kaa . server . appenders . flume . appender . FlumeAvroEventBuilder ( ) ; org . kaaproject . kaa . server . appenders . flume . config . gen . FlumeConfig flumeConfig = new org . kaaproject . kaa . server . appenders . flume . config . gen . FlumeConfig ( ) ; flumeConfig . setFlumeEventFormat ( FlumeEventFormat . RECORDS_CONTAINER ) ; flumeConfig . setIncludeClientProfile ( false ) ; flumeConfig . setIncludeServerProfile ( false ) ; eventBuilder . init ( flumeConfig ) ; eventPack = generateEventPack ( new java . util . ArrayList < org . kaaproject . kaa . server . common . log . shared . appender . LogEvent > ( ) ) ; java . util . List < org . apache . flume . Event > events = eventBuilder . generateEvents ( eventPack , null , appToken ) ; "<AssertPlaceHolder>" ; } generateEventPack ( java . util . List ) { org . kaaproject . kaa . server . common . log . shared . appender . data . BaseLogEventPack eventPack = new org . kaaproject . kaa . server . common . log . shared . appender . data . BaseLogEventPack ( profileDto , java . lang . System . currentTimeMillis ( ) , org . kaaproject . kaa . server . appenders . flume . appender . FlumeAvroEventBuilderTest . SCHEMA_VERSION , list ) ; org . kaaproject . kaa . common . dto . logs . LogSchemaDto schemaDto = new org . kaaproject . kaa . common . dto . logs . LogSchemaDto ( ) ; schemaDto . setVersion ( org . kaaproject . kaa . server . appenders . flume . appender . FlumeAvroEventBuilderTest . SCHEMA_VERSION ) ; org . kaaproject . kaa . server . common . log . shared . appender . LogSchema logSchema = new org . kaaproject . kaa . server . common . log . shared . appender . LogSchema ( schemaDto , "" ) ; eventPack . setLogSchema ( logSchema ) ; org . kaaproject . kaa . server . common . log . shared . appender . data . BaseSchemaInfo schemaInfo = new org . kaaproject . kaa . server . common . log . shared . appender . data . BaseSchemaInfo ( "1" , getResourceAsString ( org . kaaproject . kaa . server . appenders . flume . appender . FlumeAvroEventBuilderTest . SERVER_PROFILE_SCHEMA_FILE ) ) ; java . lang . String body = this . getResourceAsString ( org . kaaproject . kaa . server . appenders . flume . appender . FlumeAvroEventBuilderTest . SERVER_PROFILE_CONTENT_FILE ) ; eventPack . setServerProfile ( new org . kaaproject . kaa . server . common . log . shared . appender . data . BaseProfileInfo ( schemaInfo , body ) ) ; return eventPack ; }
org . junit . Assert . assertNull ( events )
shouldGenerateSameHashForSameInput ( ) { final long expected = this . testee . hash ( org . pitest . coverage . codeassist . ClassUtils . classAsBytes ( java . lang . String . class ) ) ; "<AssertPlaceHolder>" ; } classAsBytes ( java . lang . Class ) { return org . pitest . coverage . codeassist . ClassUtils . classAsBytes ( clazz . getName ( ) ) ; }
org . junit . Assert . assertEquals ( expected , this . testee . hash ( org . pitest . coverage . codeassist . ClassUtils . classAsBytes ( java . lang . String . class ) ) )
testCall2 ( ) { @ eu . toolchain . concurrent . SuppressWarnings ( "unchecked" ) final java . util . concurrent . Callable < java . lang . Object > callable = mock ( java . util . concurrent . Callable . class ) ; doReturn ( completable ) . when ( underTest ) . completable ( ) ; doReturn ( future ) . when ( underTest ) . doCall ( callable , executor , completable ) ; "<AssertPlaceHolder>" ; verify ( underTest ) . completable ( ) ; verify ( underTest ) . doCall ( callable , executor , completable ) ; } call ( java . util . concurrent . Callable , java . util . concurrent . ExecutorService ) { return doCall ( callable , executor , this . completable ( ) ) ; }
org . junit . Assert . assertEquals ( future , underTest . call ( callable , executor ) )
shouldIgnoreAccessControlIfTheAccessControlDisabledPropertyIsTrue ( ) { java . lang . System . setProperty ( uk . gov . justice . services . core . accesscontrol . DefaultAccessControlService . ACCESS_CONTROL_DISABLED_PROPERTY , "true" ) ; final java . util . Optional < uk . gov . justice . services . core . accesscontrol . AccessControlViolation > accessControlViolation = accessControlService . checkAccessControl ( "command" , jsonEnvelope ) ; "<AssertPlaceHolder>" ; verifyZeroInteractions ( policyEvaluator ) ; verify ( logger ) . trace ( "Skipping<sp>access<sp>control<sp>due<sp>to<sp>configuration" ) ; } checkAccessControl ( java . lang . String , uk . gov . justice . services . messaging . JsonEnvelope ) { final java . util . Optional < uk . gov . justice . services . core . accesscontrol . AccessControlViolation > accessControlViolation = accessControlService . checkAccessControl ( component , jsonEnvelope ) ; if ( accessControlViolation . isPresent ( ) ) { final java . lang . String errorMessage = accessControlFailureMessageGenerator . errorMessageFrom ( jsonEnvelope , accessControlViolation . get ( ) ) ; throw new uk . gov . justice . services . core . accesscontrol . AccessControlViolationException ( errorMessage ) ; } }
org . junit . Assert . assertThat ( accessControlViolation . isPresent ( ) , org . hamcrest . CoreMatchers . is ( false ) )
testChargeRule ( ) { org . openscience . cdk . formula . rules . IRule rule = new org . openscience . cdk . formula . rules . ChargeRule ( ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( rule )
returnsNullWhenCookieNotFoundInRequestAndDefaultValueIsNotSet ( ) { when ( cookieParam . value ( ) ) . thenReturn ( "no<sp>cookie" ) ; when ( parameter . getParameterClass ( ) ) . thenReturn ( ( ( java . lang . Class ) ( javax . ws . rs . core . Cookie . class ) ) ) ; java . lang . Object resolvedCookie = cookieParameterResolver . resolve ( parameter , applicationContext ) ; "<AssertPlaceHolder>" ; } resolve ( org . everrest . core . impl . method . org . everrest . core . Parameter , org . everrest . core . ApplicationContext ) { java . lang . String param = matrixParam . value ( ) ; org . everrest . core . method . TypeProducer typeProducer = typeProducerFactory . createTypeProducer ( parameter . getParameterClass ( ) , parameter . getGenericType ( ) ) ; java . util . List < javax . ws . rs . core . PathSegment > pathSegments = context . getUriInfo ( ) . getPathSegments ( ( ! ( parameter . isEncoded ( ) ) ) ) ; javax . ws . rs . core . PathSegment pathSegment = com . google . common . collect . Iterables . getLast ( pathSegments , null ) ; return typeProducer . createValue ( param , ( pathSegment == null ? new javax . ws . rs . core . MultivaluedHashMap ( ) : pathSegment . getMatrixParameters ( ) ) , parameter . getDefaultValue ( ) ) ; }
org . junit . Assert . assertNull ( resolvedCookie )
sort_string_array_in_java8_by_length ( ) { java . lang . String [ ] wiStateParks = new java . lang . String [ ] { "Mill<sp>Bluff<sp>State<sp>Park" , "Amnicon<sp>Falls<sp>State<sp>Park" , "Wyalusing<sp>State<sp>Park" , "Big<sp>Foot<sp>Beach<sp>State<sp>Park" , "Willow<sp>River<sp>State<sp>Park" , "Roche-A-Cri<sp>State<sp>Park" } ; java . util . Arrays . sort ( wiStateParks , ( s1 , s2 ) -> ( s1 . length ( ) ) - ( s2 . length ( ) ) ) ; com . levelup . java . array . SortArray . logger . info ( java . util . Arrays . toString ( wiStateParks ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertArrayEquals ( new java . lang . String [ ] { "Wyalusing<sp>State<sp>Park" , "Mill<sp>Bluff<sp>State<sp>Park" , "Roche-A-Cri<sp>State<sp>Park" , "Willow<sp>River<sp>State<sp>Park" , "Amnicon<sp>Falls<sp>State<sp>Park" , "Big<sp>Foot<sp>Beach<sp>State<sp>Park" } , wiStateParks )
notifyPrivateMessageArrivedWhenGuiHiddenAndPrivateChatFocusedShouldDoNothing ( ) { when ( kouChatFrame . isVisible ( ) ) . thenReturn ( false ) ; when ( kouChatFrame . isFocused ( ) ) . thenReturn ( false ) ; when ( privchat . isVisible ( ) ) . thenReturn ( true ) ; when ( privchat . isFocused ( ) ) . thenReturn ( true ) ; mediator . notifyPrivateMessageArrived ( user , null ) ; verifyZeroInteractions ( sysTray , beeper ) ; "<AssertPlaceHolder>" ; } isNewMsg ( ) { return newMsg ; }
org . junit . Assert . assertFalse ( me . isNewMsg ( ) )
getProcessDefinitionId_should_return_id_of_process_definition_with_given_name_and_version ( ) { final java . util . Map < java . lang . String , java . lang . Object > parameters = new java . util . HashMap ( ) ; java . lang . String name = "proc" ; java . lang . String version = "1.0" ; parameters . put ( "name" , name ) ; parameters . put ( "version" , version ) ; org . bonitasoft . engine . persistence . SelectOneDescriptor < java . lang . Long > selectOneDescriptor = new org . bonitasoft . engine . persistence . SelectOneDescriptor ( "getProcessDefinitionIdByNameAndVersion" , parameters , org . bonitasoft . engine . core . process . definition . model . SProcessDefinitionDeployInfo . class , org . bonitasoft . engine . core . process . definition . Long . class ) ; final long processId = 9 ; doReturn ( processId ) . when ( persistenceService ) . selectOne ( selectOneDescriptor ) ; final long result = processDefinitionServiceImpl . getProcessDefinitionId ( name , version ) ; "<AssertPlaceHolder>" ; } getProcessDefinitionId ( java . lang . String , java . lang . String ) { final org . bonitasoft . engine . service . TenantServiceAccessor tenantAccessor = getTenantAccessor ( ) ; final org . bonitasoft . engine . core . process . definition . ProcessDefinitionService processDefinitionService = tenantAccessor . getProcessDefinitionService ( ) ; try { return processDefinitionService . getProcessDefinitionId ( name , version ) ; } catch ( final org . bonitasoft . engine . core . process . definition . exception . SProcessDefinitionNotFoundException e ) { throw new org . bonitasoft . engine . bpm . process . ProcessDefinitionNotFoundException ( e ) ; } catch ( final org . bonitasoft . engine . persistence . SBonitaReadException e ) { throw new org . bonitasoft . engine . exception . RetrieveException ( e ) ; } }
org . junit . Assert . assertEquals ( processId , result )
testSubtractFromSetWithEOF ( ) { org . antlr . v4 . runtime . misc . IntervalSet s = org . antlr . v4 . runtime . misc . IntervalSet . of ( 10 , 20 ) ; s . add ( Token . EOF ) ; org . antlr . v4 . runtime . misc . IntervalSet s2 = org . antlr . v4 . runtime . misc . IntervalSet . of ( 12 , 15 ) ; java . lang . String expecting = "{<EOF>,<sp>10..11,<sp>16..20}" ; java . lang . String result = s . subtract ( s2 ) . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( a ) + ".." ) + ( b ) ; }
org . junit . Assert . assertEquals ( expecting , result )
testRemoveServiceAlert ( ) { org . onebusaway . transit_data_federation . impl . service_alerts . ServiceAlertRecord alert1 = new org . onebusaway . transit_data_federation . impl . service_alerts . ServiceAlertRecord ( ) ; alert1 . setAgencyId ( "1" ) ; alert1 . setServiceAlertId ( "A" ) ; _service . createOrUpdateServiceAlert ( alert1 ) ; _service . removeServiceAlert ( org . onebusaway . gtfs . model . AgencyAndId . convertFromString ( "1_A" ) ) ; "<AssertPlaceHolder>" ; } getServiceAlertForId ( java . lang . String ) { org . onebusaway . gtfs . model . AgencyAndId id = org . onebusaway . transit_data_federation . impl . federated . AgencyAndIdLibrary . convertFromString ( situationId ) ; return _serviceAlertsBeanService . getServiceAlertForId ( id ) ; }
org . junit . Assert . assertNull ( _service . getServiceAlertForId ( org . onebusaway . gtfs . model . AgencyAndId . convertFromString ( "1_A" ) ) )
paintingTest ( ) { java . io . ByteArrayOutputStream byteOutput = new java . io . ByteArrayOutputStream ( ) ; java . io . DataOutputStream outStream = new java . io . DataOutputStream ( byteOutput ) ; org . logblock . entry . blob . PaintingBlob blobOut = org . logblock . entry . BlobEntry . create ( 1 , org . logblock . entry . blob . PaintingBlob . class ) ; blobOut . setArt ( "artistic" ) ; blobOut . setDirection ( ( ( byte ) ( 5 ) ) ) ; blobOut . write ( outStream ) ; outStream . close ( ) ; java . io . DataInputStream inputStream = new java . io . DataInputStream ( new java . io . ByteArrayInputStream ( byteOutput . toByteArray ( ) ) ) ; org . logblock . entry . blob . PaintingBlob blobIn = org . logblock . entry . BlobEntry . create ( 1 , org . logblock . entry . blob . PaintingBlob . class ) ; blobIn . read ( inputStream ) ; "<AssertPlaceHolder>" ; } read ( java . io . DataInput ) { art = in . readUTF ( ) ; direction = in . readByte ( ) ; }
org . junit . Assert . assertEquals ( blobOut , blobIn )
setSmallerTtlForDirectoryWithTtl ( ) { alluxio . master . file . contexts . CreateDirectoryContext directoryContext = alluxio . master . file . contexts . CreateDirectoryContext . mergeFrom ( alluxio . grpc . CreateDirectoryPOptions . newBuilder ( ) . setCommonOptions ( alluxio . grpc . FileSystemMasterCommonPOptions . newBuilder ( ) . setTtl ( Constants . HOUR_MS ) ) . setRecursive ( true ) ) ; mFileSystemMaster . createDirectory ( alluxio . master . file . FileSystemMasterTest . NESTED_URI , directoryContext ) ; alluxio . heartbeat . HeartbeatScheduler . execute ( HeartbeatContext . MASTER_TTL_CHECK ) ; "<AssertPlaceHolder>" ; mFileSystemMaster . setAttribute ( alluxio . master . file . FileSystemMasterTest . NESTED_URI , alluxio . master . file . contexts . SetAttributeContext . mergeFrom ( alluxio . grpc . SetAttributePOptions . newBuilder ( ) . setCommonOptions ( alluxio . grpc . FileSystemMasterCommonPOptions . newBuilder ( ) . setTtl ( 0 ) ) ) ) ; alluxio . heartbeat . HeartbeatScheduler . execute ( HeartbeatContext . MASTER_TTL_CHECK ) ; mThrown . expect ( alluxio . exception . FileDoesNotExistException . class ) ; mFileSystemMaster . getFileInfo ( alluxio . master . file . FileSystemMasterTest . NESTED_URI , alluxio . master . file . FileSystemMasterTest . GET_STATUS_CONTEXT ) ; } getFileInfo ( alluxio . grpc . GetFileInfoPRequest , io . grpc . stub . StreamObserver ) { final long fileId = request . getFileId ( ) ; alluxio . grpc . GetFileInfoPOptions options = request . getOptions ( ) ; alluxio . RpcUtils . call ( alluxio . master . file . FileSystemMasterWorkerServiceHandler . LOG , ( ( alluxio . RpcUtils . RpcCallableThrowsIOException < alluxio . grpc . GetFileInfoPResponse > ) ( ( ) -> alluxio . grpc . GetFileInfoPResponse . newBuilder ( ) . setFileInfo ( alluxio . grpc . GrpcUtils . toProto ( mFileSystemMaster . getFileInfo ( fileId ) ) ) . build ( ) ) ) , "getFileInfo" , "fileId=%s,<sp>options=%s" , responseObserver , fileId , options ) ; }
org . junit . Assert . assertTrue ( ( ( mFileSystemMaster . getFileInfo ( alluxio . master . file . FileSystemMasterTest . NESTED_URI , alluxio . master . file . FileSystemMasterTest . GET_STATUS_CONTEXT ) . getName ( ) ) != null ) )
testListAllWithLimit ( ) { java . util . List < uk . ac . bbsrc . tgac . miso . core . data . Study > studies = dao . listAllWithLimit ( 2L ) ; "<AssertPlaceHolder>" ; } size ( ) { return map . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , studies . size ( ) )
testGetSessionId ( ) { final java . lang . String clientId = "1" ; final java . lang . String sessionId = "2" ; when ( request . getSession ( eq ( true ) ) ) . thenReturn ( httpSession ) ; when ( request . getParameter ( eq ( "clientId" ) ) ) . thenReturn ( clientId ) ; when ( provider . createOrGetSession ( httpSession , clientId ) ) . thenReturn ( queueSession ) ; when ( queueSession . getSessionId ( ) ) . thenReturn ( sessionId ) ; "<AssertPlaceHolder>" ; } abstractFileServlet ( ) { return new org . guvnor . common . services . backend . file . upload . AbstractFileServlet ( ) { @ org . guvnor . common . services . backend . file . upload . Override protected java . io . InputStream doLoad ( final org . uberfire . backend . vfs . Path path , final javax . servlet . http . HttpServletRequest request ) { return null ; } @ org . guvnor . common . services . backend . file . upload . Override protected void doCreate ( final org . uberfire . backend . vfs . Path path , final java . io . InputStream data , final javax . servlet . http . HttpServletRequest request , final java . lang . String comment ) { } @ org . guvnor . common . services . backend . file . upload . Override protected void doUpdate ( final org . uberfire . backend . vfs . Path path , final java . io . InputStream data , final javax . servlet . http . HttpServletRequest request , final java . lang . String comment ) { } @ org . guvnor . common . services . backend . file . upload . Override protected org . uberfire . backend . vfs . Path convertPath ( final java . lang . String fileName , final java . lang . String contextPath ) throws java . net . URISyntaxException { return null ; } @ org . guvnor . common . services . backend . file . upload . Override protected org . uberfire . backend . vfs . Path convertPath ( final java . lang . String fullPath ) throws java . net . URISyntaxException { return null ; } } ; }
org . junit . Assert . assertEquals ( sessionId , abstractFileServlet ( ) . getSessionId ( request , provider ) )
testCreateRelationshipWithCommits ( ) { org . neo4j . graphdb . Node n1 = getGraphDb ( ) . createNode ( ) ; newTransaction ( ) ; n1 = getGraphDb ( ) . getNodeById ( n1 . getId ( ) ) ; org . neo4j . graphdb . Node n2 = getGraphDb ( ) . createNode ( ) ; n1 . createRelationshipTo ( n2 , org . neo4j . kernel . impl . MyRelTypes . TEST ) ; newTransaction ( ) ; org . neo4j . graphdb . Relationship [ ] relArray = getRelationshipArray ( n1 . getRelationships ( ) ) ; "<AssertPlaceHolder>" ; relArray = getRelationshipArray ( n1 . getRelationships ( ) ) ; relArray [ 0 ] . delete ( ) ; n1 . delete ( ) ; n2 . delete ( ) ; } getRelationships ( ) { return null ; }
org . junit . Assert . assertEquals ( 1 , relArray . length )
getsAndSetsStartOffset ( ) { com . mpatric . mp3agic . ID3v2ChapterFrameData frameData = new com . mpatric . mp3agic . ID3v2ChapterFrameData ( false ) ; frameData . setStartOffset ( 9 ) ; "<AssertPlaceHolder>" ; } getStartOffset ( ) { return startOffset ; }
org . junit . Assert . assertEquals ( 9 , frameData . getStartOffset ( ) )
testAllNulls ( ) { System . out . println ( "test<sp>all<sp>nulls<sp>============================================================================" ) ; java . util . Map < org . apache . kylin . metadata . model . TblColRef , java . util . Set < java . lang . String > > values = com . google . common . collect . Maps . newHashMap ( ) ; values . put ( org . apache . kylin . storage . hbase . FuzzyValueCombinationTest . col1 , set ( ) ) ; values . put ( org . apache . kylin . storage . hbase . FuzzyValueCombinationTest . col2 , set ( ) ) ; values . put ( org . apache . kylin . storage . hbase . FuzzyValueCombinationTest . col3 , set ( ) ) ; java . util . List < java . util . Map < org . apache . kylin . metadata . model . TblColRef , java . lang . String > > result = org . apache . kylin . storage . hbase . FuzzyValueCombination . calculate ( values , 10 ) ; for ( java . util . Map < org . apache . kylin . metadata . model . TblColRef , java . lang . String > item : result ) { System . out . println ( item ) ; } "<AssertPlaceHolder>" ; } size ( ) { return columns . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , result . size ( ) )
testBasicJobPlanSerde ( ) { org . apache . tez . dag . api . records . DAGProtos . DAGPlan job = org . apache . tez . dag . api . records . DAGProtos . DAGPlan . newBuilder ( ) . setName ( "test" ) . addVertex ( org . apache . tez . dag . api . records . DAGProtos . VertexPlan . newBuilder ( ) . setName ( "vertex1" ) . setType ( PlanVertexType . NORMAL ) . addTaskLocationHint ( org . apache . tez . dag . api . records . DAGProtos . PlanTaskLocationHint . newBuilder ( ) . addHost ( "machineName" ) . addRack ( "rack1" ) . build ( ) ) . setTaskConfig ( org . apache . tez . dag . api . records . DAGProtos . PlanTaskConfiguration . newBuilder ( ) . setNumTasks ( 2 ) . setVirtualCores ( 4 ) . setMemoryMb ( 1024 ) . setJavaOpts ( "" ) . setTaskModule ( "x.y" ) . build ( ) ) . build ( ) ) . build ( ) ; java . io . File file = tempFolder . newFile ( "jobPlan" ) ; java . io . FileOutputStream outStream = null ; try { outStream = new java . io . FileOutputStream ( file ) ; job . writeTo ( outStream ) ; } finally { if ( outStream != null ) { outStream . close ( ) ; } } org . apache . tez . dag . api . records . DAGProtos . DAGPlan inJob ; java . io . FileInputStream inputStream ; try { inputStream = new java . io . FileInputStream ( file ) ; inJob = org . apache . tez . dag . api . records . DAGProtos . DAGPlan . newBuilder ( ) . mergeFrom ( inputStream ) . build ( ) ; } finally { outStream . close ( ) ; } "<AssertPlaceHolder>" ; } close ( ) { if ( ( curReader ) != null ) { curReader . close ( ) ; curReader = null ; } }
org . junit . Assert . assertEquals ( job , inJob )
testGetPath ( ) { final java . lang . String path = "path" ; final org . apache . oozie . fluentjob . api . action . Touchz touchz = new org . apache . oozie . fluentjob . api . action . Touchz ( path ) ; "<AssertPlaceHolder>" ; } getPath ( ) { return path ; }
org . junit . Assert . assertEquals ( path , touchz . getPath ( ) )
testGetFile ( ) { org . apache . cxf . tools . util . FileWriterUtil fileWriter = null ; java . lang . String tmpDir = java . lang . System . getProperty ( "java.io.tmpdir" ) ; java . io . File targetDir = new java . io . File ( ( ( tmpDir + ( java . io . File . separator ) ) + "target" ) ) ; try { targetDir . mkdirs ( ) ; fileWriter = new org . apache . cxf . tools . util . FileWriterUtil ( targetDir . getAbsolutePath ( ) , null ) ; fileWriter . getWriter ( "com.iona.test" , "A.java" ) ; java . lang . String packPath = "/com/iona/test/A.java" . replace ( '/' , File . separatorChar ) ; java . lang . String path = ( targetDir . getAbsolutePath ( ) ) + packPath ; "<AssertPlaceHolder>" ; } catch ( java . io . IOException e ) { e . printStackTrace ( ) ; } finally { cleanDir ( targetDir ) ; } } getName ( ) { return name ; }
org . junit . Assert . assertNotNull ( new java . io . File ( path ) . getName ( ) )
testSqlDate ( ) { java . lang . Class < com . j256 . ormlite . field . types . SqlDateTypeTest . LocalDate > clazz = com . j256 . ormlite . field . types . SqlDateTypeTest . LocalDate . class ; com . j256 . ormlite . dao . Dao < com . j256 . ormlite . field . types . SqlDateTypeTest . LocalDate , java . lang . Object > dao = createDao ( clazz , true ) ; java . util . GregorianCalendar c = new java . util . GregorianCalendar ( ) ; c . set ( GregorianCalendar . HOUR_OF_DAY , 0 ) ; c . set ( GregorianCalendar . MINUTE , 0 ) ; c . set ( GregorianCalendar . SECOND , 0 ) ; c . set ( GregorianCalendar . MILLISECOND , 0 ) ; long millis = c . getTimeInMillis ( ) ; java . sql . Date val = new java . sql . Date ( millis ) ; java . lang . String format = "yyyy-MM-dd<sp>HH:mm:ss.S" ; java . text . DateFormat dateFormat = new java . text . SimpleDateFormat ( format ) ; java . lang . String valStr = dateFormat . format ( val ) ; com . j256 . ormlite . field . types . SqlDateTypeTest . LocalDate foo = new com . j256 . ormlite . field . types . SqlDateTypeTest . LocalDate ( ) ; foo . date = val ; "<AssertPlaceHolder>" ; java . sql . Timestamp timestamp = new java . sql . Timestamp ( val . getTime ( ) ) ; testType ( dao , foo , clazz , val , timestamp , timestamp , valStr , dataType , com . j256 . ormlite . field . types . SqlDateTypeTest . DATE_COLUMN , false , true , true , false , true , false , true , false ) ; } create ( T ) { checkForInitialized ( ) ; if ( data == null ) { return 0 ; } if ( data instanceof com . j256 . ormlite . misc . BaseDaoEnabled ) { @ com . j256 . ormlite . dao . SuppressWarnings ( "unchecked" ) com . j256 . ormlite . misc . BaseDaoEnabled < T , ID > daoEnabled = ( ( com . j256 . ormlite . misc . BaseDaoEnabled < T , ID > ) ( data ) ) ; daoEnabled . setDao ( this ) ; } com . j256 . ormlite . support . DatabaseConnection connection = connectionSource . getReadWriteConnection ( tableInfo . getTableName ( ) ) ; try { return statementExecutor . create ( connection , data , objectCache ) ; } finally { connectionSource . releaseConnection ( connection ) ; } }
org . junit . Assert . assertEquals ( 1 , dao . create ( foo ) )
testGetKeySemaphoresOfPartitioningResourceIds ( ) { org . hivedb . Hive hive = org . hivedb . Hive . load ( getConnectString ( H2TestCase . TEST_DB ) , org . hivedb . meta . persistence . CachingDataSourceProvider . getInstance ( ) ) ; hive . deleteResource ( resource ) ; resource = org . hivedb . util . functional . Atom . getFirstOrNull ( dimension . getResources ( ) ) ; resource . setIsPartitioningResource ( true ) ; hive . addResource ( resource ) ; resource = hive . getPartitionDimension ( ) . getResource ( resource . getName ( ) ) ; insertKeys ( getHive ( ) ) ; org . hivedb . meta . directory . DbDirectory d = getDirectory ( ) ; for ( java . lang . String key : getPrimaryIndexOrResourceKeys ( ) ) "<AssertPlaceHolder>" ; } getKeySemaphoresOfResourceId ( org . hivedb . meta . Resource , java . lang . Object ) { return ( ( java . util . Collection < org . hivedb . meta . directory . KeySemaphore > ) ( resource . isPartitioningResource ( ) ? getKeySemamphoresOfPrimaryIndexKey ( resourceId ) : doRead ( sql . selectKeySemaphoresOfResourceId ( resource ) , new java . lang . Object [ ] { resourceId } , new org . hivedb . meta . directory . KeySemaphoreRowMapper ( ) ) ) ) ; }
org . junit . Assert . assertEquals ( 1 , d . getKeySemaphoresOfResourceId ( resource , key ) . size ( ) )
testEC2SetupMock ( ) { org . oscm . app . aws . EC2Communication . useMock ( ec2 ) ; org . oscm . app . aws . EC2Communication realEC2 = new org . oscm . app . aws . EC2Communication ( ph ) ; com . amazonaws . services . ec2 . AmazonEC2 client = realEC2 . getEC2 ( ) ; "<AssertPlaceHolder>" ; } getEC2 ( ) { if ( ( ec2 ) == null ) { java . lang . String endpoint = ( ( org . oscm . app . aws . EC2Communication . ENDPOINT_PREFIX ) + ( ph . getRegion ( ) ) ) + ( org . oscm . app . aws . EC2Communication . ENDPOINT_SUFFIX ) ; java . lang . String proxyHost = java . lang . System . getProperty ( org . oscm . app . aws . EC2Communication . HTTPS_PROXY_HOST ) ; java . lang . String proxyPort = java . lang . System . getProperty ( org . oscm . app . aws . EC2Communication . HTTPS_PROXY_PORT ) ; java . lang . String proxyUser = java . lang . System . getProperty ( org . oscm . app . aws . EC2Communication . HTTPS_PROXY_USER ) ; java . lang . String proxyPassword = java . lang . System . getProperty ( org . oscm . app . aws . EC2Communication . HTTPS_PROXY_PASSWORD ) ; int proxyPortInt = 0 ; try { proxyPortInt = java . lang . Integer . parseInt ( proxyPort ) ; } catch ( java . lang . NumberFormatException e ) { } com . amazonaws . ClientConfiguration clientConfiguration = new com . amazonaws . ClientConfiguration ( ) ; if ( ! ( isNonProxySet ( endpoint ) ) ) { if ( proxyHost != null ) { clientConfiguration . setProxyHost ( proxyHost ) ; } if ( proxyPortInt > 0 ) { clientConfiguration . setProxyPort ( proxyPortInt ) ; } if ( ( proxyUser != null ) && ( ( proxyUser . length ( ) ) > 0 ) ) { clientConfiguration . setProxyUsername ( proxyUser ) ; } if ( ( proxyPassword != null ) && ( ( proxyPassword . length ( ) ) > 0 ) ) { clientConfiguration . setProxyPassword ( proxyPassword ) ; } } ec2 = getEC2 ( credentialsProvider , clientConfiguration ) ; ec2 . setEndpoint ( endpoint ) ; } return ec2 ; }
org . junit . Assert . assertTrue ( ( client == ( ec2 ) ) )
testJiraCsv154_withHeaderComments ( ) { final java . lang . String comment = "This<sp>is<sp>a<sp>header<sp>comment" ; final org . apache . commons . csv . CSVFormat format = CSVFormat . EXCEL . withHeader ( "H1" , "H2" ) . withHeaderComments ( comment ) . withCommentMarker ( '#' ) ; final java . lang . StringBuilder out = new java . lang . StringBuilder ( ) ; try ( final org . apache . commons . csv . CSVPrinter printer = format . print ( out ) ) { printer . print ( "A" ) ; printer . print ( "B" ) ; } final java . lang . String s = out . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( ( ( "CSVRecord<sp>[comment=" + ( comment ) ) + ",<sp>mapping=" ) + ( mapping ) ) + ",<sp>recordNumber=" ) + ( recordNumber ) ) + ",<sp>values=" ) + ( java . util . Arrays . toString ( values ) ) ) + "]" ; }
org . junit . Assert . assertTrue ( s , s . contains ( comment ) )
convertLogServerSyncTest ( ) { org . kaaproject . kaa . server . sync . platform . AvroEncDec avroEncDec = new org . kaaproject . kaa . server . sync . platform . AvroEncDec ( ) ; java . lang . reflect . Method method = org . kaaproject . kaa . server . sync . platform . AvroEncDec . class . getDeclaredMethod ( org . kaaproject . kaa . server . sync . platform . AvroEncDecTest . CONVER_METHOD , org . kaaproject . kaa . server . sync . LogServerSync . class ) ; method . setAccessible ( true ) ; org . kaaproject . kaa . server . sync . LogServerSync serverSync = null ; "<AssertPlaceHolder>" ; method . invoke ( avroEncDec , new org . kaaproject . kaa . server . sync . LogServerSync ( ) ) ; }
org . junit . Assert . assertNull ( method . invoke ( avroEncDec , serverSync ) )
testLocationChangeFromNullToFoobar ( ) { testLocationChanges ( "classpath:/import_dummy-null.xml" , "Default" , "classpath:/import_dummy-foobar.xml" , "foobar" ) ; java . util . List < org . opennms . netmgt . model . monitoringLocations . OnmsMonitoringLocation > locations = m_locationDao . findAll ( ) ; "<AssertPlaceHolder>" ; } size ( ) { m_lock . readLock ( ) . lock ( ) ; try { return m_managedAddresses . size ( ) ; } finally { m_lock . readLock ( ) . unlock ( ) ; } }
org . junit . Assert . assertEquals ( 2 , locations . size ( ) )
zouMeldingMoetenGevenOmdatDatumAanvangOverlijdenInToekomstLigt ( ) { final java . util . List < nl . bzk . brp . model . basis . BerichtEntiteit > meldingen = brby0011 . voerRegelUit ( null , null , nl . bzk . brp . bijhouding . business . regels . util . ActieBerichtBuilder . bouwNieuweActie ( SoortActie . REGISTRATIE_OVERLIJDEN ) . setDatumAanvang ( new nl . bzk . brp . model . algemeen . attribuuttype . kern . DatumEvtDeelsOnbekendAttribuut ( nl . bzk . brp . model . algemeen . attribuuttype . kern . DatumAttribuut . morgen ( ) ) ) . getActie ( ) , null ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , meldingen . size ( ) )
shouldCountCustomMethodWithWithinPoint ( ) { java . lang . String documentId = randomNumeric ( 5 ) ; org . springframework . data . elasticsearch . entities . SampleEntity sampleEntity = new org . springframework . data . elasticsearch . entities . SampleEntity ( ) ; sampleEntity . setId ( documentId ) ; sampleEntity . setType ( "test" ) ; sampleEntity . setRate ( 10 ) ; sampleEntity . setMessage ( "foo" ) ; sampleEntity . setLocation ( new org . springframework . data . elasticsearch . core . geo . GeoPoint ( 45.7806 , 3.0875 ) ) ; repository . save ( sampleEntity ) ; documentId = randomNumeric ( 5 ) ; org . springframework . data . elasticsearch . entities . SampleEntity sampleEntity2 = new org . springframework . data . elasticsearch . entities . SampleEntity ( ) ; sampleEntity2 . setId ( documentId ) ; sampleEntity2 . setType ( "test" ) ; sampleEntity2 . setRate ( 10 ) ; sampleEntity2 . setMessage ( "foo" ) ; sampleEntity2 . setLocation ( new org . springframework . data . elasticsearch . core . geo . GeoPoint ( 30.7806 , 0.0875 ) ) ; repository . save ( sampleEntity2 ) ; long count = repository . countByLocationWithin ( new org . springframework . data . geo . Point ( 45.7806 , 3.0875 ) , new org . springframework . data . geo . Distance ( 2 , org . springframework . data . geo . Metrics . KILOMETERS ) ) ; "<AssertPlaceHolder>" ; } is ( java . lang . Object ) { queryCriteria . add ( new org . springframework . data . elasticsearch . core . query . Criteria . CriteriaEntry ( org . springframework . data . elasticsearch . core . query . Criteria . OperationKey . EQUALS , o ) ) ; return this ; }
org . junit . Assert . assertThat ( count , is ( equalTo ( 1L ) ) )