input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
getPartsShouldReturnCurrentPart ( ) { "<AssertPlaceHolder>" ; } getParts ( ) { return java . util . Collections . unmodifiableSet ( partTabIndex . keySet ( ) ) ; }
org . junit . Assert . assertFalse ( view . getParts ( ) . isEmpty ( ) )
testEquals ( ) { org . talend . designer . xmlmap . model . emf . xmlmap . XmlMapData data1 = XmlmapFactory . eINSTANCE . createXmlMapData ( ) ; org . talend . designer . xmlmap . model . emf . xmlmap . XmlMapData data2 = XmlmapFactory . eINSTANCE . createXmlMapData ( ) ; "<AssertPlaceHolder>" ; test1 ( ) ; test2 ( ) ; test3 ( ) ; test4 ( ) ; } equals ( java . lang . Object ) { if ( o == ( this ) ) return true ; if ( ! ( o instanceof org . talend . sdk . component . studio . metadata . node . TaCoKitFamilyRepositoryNode ) ) return false ; final org . talend . sdk . component . studio . metadata . node . TaCoKitFamilyRepositoryNode other = ( ( org . talend . sdk . component . studio . metadata . node . TaCoKitFamilyRepositoryNode ) ( o ) ) ; if ( ! ( other . canEqual ( this ) ) ) return false ; if ( ! ( super . equals ( o ) ) ) return false ; final java . lang . Object this$configTypeNode = this . getConfigTypeNode ( ) ; final java . lang . Object other$configTypeNode = other . getConfigTypeNode ( ) ; if ( this$configTypeNode == null ? other$configTypeNode != null : ! ( this$configTypeNode . equals ( other$configTypeNode ) ) ) return false ; return true ; }
org . junit . Assert . assertTrue ( data1 . equals ( data2 ) )
shouldCoerceATargetWithASingleFlavor ( ) { com . facebook . buck . core . model . BuildTarget seen = new com . facebook . buck . rules . coercer . BuildTargetTypeCoercer ( unconfiguredBuildTargetTypeCoercer ) . coerce ( createCellRoots ( filesystem ) , filesystem , basePath , EmptyTargetConfiguration . INSTANCE , "//foo:bar#baz" ) ; "<AssertPlaceHolder>" ; } newInstance ( java . lang . String ) { return com . facebook . buck . core . model . BuildTargetFactory . newInstance ( ( ( java . nio . file . Path ) ( null ) ) , fullyQualifiedName ) ; }
org . junit . Assert . assertEquals ( com . facebook . buck . core . model . BuildTargetFactory . newInstance ( "//foo:bar#baz" ) , seen )
testIntersectionWithTwoContainedElementsReversed ( ) { org . antlr . misc . IntervalSet s = org . antlr . misc . IntervalSet . of ( 10 , 20 ) ; org . antlr . misc . IntervalSet s2 = org . antlr . misc . IntervalSet . of ( 2 , 2 ) ; s2 . add ( 15 ) ; s2 . add ( 18 ) ; java . lang . String expecting = "{15,<sp>18}" ; java . lang . String result = s2 . and ( s ) . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return new java . lang . String ( data ) ; }
org . junit . Assert . assertEquals ( result , expecting )
testRemoveValidatorLatitudeNonExist ( ) { "<AssertPlaceHolder>" ; } removeValidator ( com . tech . configurations . tools . ValidationScopes , int ) { if ( i == 0 ) { return false ; } switch ( scope ) { case LONGITUDE : if ( ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LONGITUDE_VALIDATORS . size ( ) ) >= ( i + 1 ) ) { com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LONGITUDE_VALIDATORS . get ( ( i - 1 ) ) . replaceNext ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LONGITUDE_VALIDATORS . get ( i ) . getNext ( ) ) ; com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LONGITUDE_VALIDATORS . remove ( i ) ; return true ; } return false ; case LATITUDE : if ( ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LATITUDE_VALIDATORS . size ( ) ) >= ( i + 1 ) ) { com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LATITUDE_VALIDATORS . get ( ( i - 1 ) ) . replaceNext ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LATITUDE_VALIDATORS . get ( i ) . getNext ( ) ) ; com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . LATITUDE_VALIDATORS . remove ( i ) ; return true ; } return false ; case ROOM_NAME : if ( ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . ROOM_NAME_VALIDATORS . size ( ) ) >= ( i + 1 ) ) { com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . ROOM_NAME_VALIDATORS . get ( ( i - 1 ) ) . replaceNext ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . ROOM_NAME_VALIDATORS . get ( i ) . getNext ( ) ) ; com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . ROOM_NAME_VALIDATORS . remove ( i ) ; return true ; } return false ; case USER_NAME : if ( ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . USER_NAME_VALIDATORS . size ( ) ) >= ( i + 1 ) ) { com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . USER_NAME_VALIDATORS . get ( ( i - 1 ) ) . replaceNext ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . USER_NAME_VALIDATORS . get ( i ) . getNext ( ) ) ; com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . USER_NAME_VALIDATORS . remove ( i ) ; return true ; } return false ; default : throw new com . tech . exceptions . customexceptions . ValidatorNotListedException ( ) ; } }
org . junit . Assert . assertFalse ( com . tech . models . dtos . chatroom . ChatroomCheckInsideDTO . removeValidator ( ValidationScopes . LATITUDE , 20 ) )
testCreate_modifires ( ) { com . github . mygreen . supercsv . builder . BeanMapping < com . github . mygreen . supercsv . builder . BeanMappingFactoryTest . AllModifiresBean > beanMapping = factory . create ( com . github . mygreen . supercsv . builder . BeanMappingFactoryTest . AllModifiresBean . class , groupEmpty ) ; "<AssertPlaceHolder>" . hasSize ( 4 ) ; } getColumns ( ) { return columns ; }
org . junit . Assert . assertThat ( beanMapping . getColumns ( ) )
testSomeMethod ( ) { byte [ ] testArray = org . spf4j . base . Base64Test . generateTestArray ( 2048 ) ; java . lang . String encodeBase64 = org . spf4j . base . Base64 . encodeBase64 ( testArray ) ; byte [ ] parseBase64 = org . spf4j . base . Base64 . decodeBase64 ( encodeBase64 ) ; "<AssertPlaceHolder>" ; } decodeBase64 ( java . lang . CharSequence ) { return org . spf4j . base . Base64 . decodeBase64 ( text , 0 , text . length ( ) ) ; }
org . junit . Assert . assertArrayEquals ( testArray , parseBase64 )
testCreateConnectionCallUnknwonUser ( ) { org . apache . qpid . jms . JmsConnectionFactory factory = new org . apache . qpid . jms . JmsConnectionFactory ( getBrokerAmqpConnectionURI ( ) ) ; connection = factory . createQueueConnection ( "unknown" , "unknown" ) ; "<AssertPlaceHolder>" ; connection . start ( ) ; } createQueueConnection ( java . lang . String , java . lang . String ) { org . apache . qpid . jms . JmsQueueConnection connection = null ; try { org . apache . qpid . jms . meta . JmsConnectionInfo connectionInfo = configureConnectionInfo ( username , password ) ; org . apache . qpid . jms . provider . Provider provider = createProvider ( remoteURI ) ; connection = new org . apache . qpid . jms . JmsQueueConnection ( connectionInfo , provider ) ; connection . setExceptionListener ( exceptionListener ) ; connection . connect ( ) ; } catch ( java . lang . Exception e ) { if ( connection != null ) { try { connection . close ( ) ; } catch ( java . lang . Throwable ignored ) { } } throw org . apache . qpid . jms . exceptions . JmsExceptionSupport . create ( e ) ; } return connection ; }
org . junit . Assert . assertNotNull ( connection )
testCheckAndMutateWithBuilder ( ) { try ( org . apache . hadoop . hbase . client . Table table = createTable ( ) ) { putOneRow ( table ) ; getOneRowAndAssertAllExist ( table ) ; org . apache . hadoop . hbase . client . RowMutations rm = makeRowMutationsWithColumnCDeleted ( ) ; boolean res = table . checkAndMutate ( org . apache . hadoop . hbase . client . TestCheckAndMutate . ROWKEY , org . apache . hadoop . hbase . client . TestCheckAndMutate . FAMILY ) . qualifier ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "A" ) ) . ifEquals ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "a" ) ) . thenMutate ( rm ) ; "<AssertPlaceHolder>" ; getOneRowAndAssertAllButCExist ( table ) ; try { rm = getBogusRowMutations ( ) ; table . checkAndMutate ( org . apache . hadoop . hbase . client . TestCheckAndMutate . ROWKEY , org . apache . hadoop . hbase . client . TestCheckAndMutate . FAMILY ) . qualifier ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "A" ) ) . ifEquals ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "a" ) ) . thenMutate ( rm ) ; org . junit . Assert . fail ( "Expected<sp>NoSuchColumnFamilyException" ) ; } catch ( org . apache . hadoop . hbase . client . RetriesExhaustedWithDetailsException e ) { try { throw e . getCause ( 0 ) ; } catch ( org . apache . hadoop . hbase . regionserver . NoSuchColumnFamilyException e1 ) { } } } } thenMutate ( org . apache . hadoop . hbase . client . RowMutations ) { preCheck ( ) ; return doCheckAndMutate ( row , family , qualifier , op . name ( ) , value , timeRange , mutation ) ; }
org . junit . Assert . assertTrue ( res )
validIntegerGreaterThanConstraintShouldNotCreateViolations ( ) { java . util . Set < javax . validation . ConstraintViolation < org . alien4cloud . tosca . model . definitions . PropertyDefinition > > violations = validator . validate ( createGreaterThanDefinition ( ToscaTypes . INTEGER . toString ( ) , 42 ) , alien4cloud . tosca . container . validation . ToscaSequence . class ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( "name:<sp>[" + ( name ) ) + "],<sp>value:<sp>[" ) + ( value ) ) + "]" ; }
org . junit . Assert . assertEquals ( 0 , violations . size ( ) )
testEmpty ( ) { formatter = new org . apache . accumulo . shell . format . DeleterFormatter ( writer , java . util . Collections . < org . apache . accumulo . core . data . Key , org . apache . accumulo . core . data . Value > emptyMap ( ) . entrySet ( ) , new org . apache . accumulo . core . util . format . FormatterConfig ( ) . setPrintTimestamps ( true ) , shellState , true ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { return iterator . hasNext ( ) ; }
org . junit . Assert . assertFalse ( formatter . hasNext ( ) )
testRecordinality ( ) { final int numRuns = 10000 ; final int [ ] kSizes = new int [ ] { 4 , 8 , 16 , 32 , 64 , 128 , 256 , 512 } ; final java . lang . String msnd = "/midsummer-nights-dream-gutenberg.txt" ; final java . net . URI uri = getClass ( ) . getResource ( msnd ) . toURI ( ) ; final java . util . List < java . lang . String > input = org . apache . commons . io . FileUtils . readLines ( new java . io . File ( uri ) ) ; final java . util . List < java . lang . String > words = com . google . common . collect . Lists . newArrayList ( ) ; for ( java . lang . String line : input ) words . add ( line . trim ( ) . toLowerCase ( ) ) ; final java . util . Set < java . lang . String > actualSet = new java . util . HashSet < java . lang . String > ( ) ; for ( java . lang . String line : words ) actualSet . add ( line ) ; final double actualSize = actualSet . size ( ) ; System . out . println ( ( "Actual<sp>cardinality:<sp>" + ( actualSet . size ( ) ) ) ) ; final com . cscotta . recordinality . ExecutorService exec = com . cscotta . recordinality . Executors . newFixedThreadPool ( java . lang . Runtime . getRuntime ( ) . availableProcessors ( ) ) ; final java . util . List < com . cscotta . recordinality . Future < com . cscotta . recordinality . RecordinalityTest . Result > > resultFutures = com . google . common . collect . Lists . newArrayList ( ) ; for ( int kSize : kSizes ) { resultFutures . add ( exec . submit ( buildRun ( kSize , numRuns , words ) ) ) ; } final double [ ] errorBounds = new double [ ] { 2.5 , 1 , 0.6 , 0.4 , 0.25 , 0.2 , 0.1 , 0.05 } ; int errorBoundIdx = 0 ; for ( com . cscotta . recordinality . Future < com . cscotta . recordinality . RecordinalityTest . Result > future : resultFutures ) { com . cscotta . recordinality . RecordinalityTest . Result result = future . get ( ) ; System . out . println ( ( "k-size:<sp>" + ( result . kSize ) ) ) ; System . out . println ( ( "Mean<sp>est:<sp>" + ( result . mean ) ) ) ; System . out . println ( ( "Std.<sp>err:<sp>" + ( result . stdError ) ) ) ; System . out . println ( ( ( "Mean<sp>run<sp>time:<sp>" + ( ( result . runTime ) / numRuns ) ) + "ms." ) ) ; System . out . println ( ( "Mean<sp>estimate<sp>is<sp>off<sp>by<sp>a<sp>factor<sp>of<sp>" + ( ( ( result . mean ) / actualSize ) - 1 ) ) ) ; System . out . println ( "\n==================================================\n" ) ; "<AssertPlaceHolder>" ; errorBoundIdx += 1 ; } exec . shutdown ( ) ; } buildRun ( int , int , java . util . List ) { return new com . cscotta . recordinality . Callable < com . cscotta . recordinality . RecordinalityTest . Result > ( ) { public com . cscotta . recordinality . RecordinalityTest . Result call ( ) throws com . cscotta . recordinality . Exception { long start = java . lang . System . currentTimeMillis ( ) ; final double [ ] results = new double [ numRuns ] ; for ( int i = 0 ; i < numRuns ; i ++ ) { com . cscotta . recordinality . Recordinality rec = new com . cscotta . recordinality . Recordinality ( kSize ) ; for ( java . lang . String line : lines ) rec . observe ( line ) ; results [ i ] = rec . estimateCardinality ( ) ; } double mean = new org . apache . commons . math3 . stat . descriptive . moment . Mean ( ) . evaluate ( results ) ; double stdDev = new org . apache . commons . math3 . stat . descriptive . moment . StandardDeviation ( ) . evaluate ( results ) ; double stdError = stdDev / 3193 ; long runTime = ( java . lang . System . currentTimeMillis ( ) ) - start ; return new com . cscotta . recordinality . RecordinalityTest . Result ( kSize , mean , stdError , runTime ) ; } } ; }
org . junit . Assert . assertTrue ( ( ( result . stdError ) <= ( errorBounds [ errorBoundIdx ] ) ) )
testRetryMaxAttemptsFileAuditDisabled ( ) { datawave . microservice . audit . AuditServiceTest . isFileAuditEnabled = false ; int maxAttempts = 2 ; long backoffIntervalMillis = 50L ; datawave . microservice . audit . config . AuditProperties . Retry retry = new datawave . microservice . audit . config . AuditProperties . Retry ( ) ; retry . setMaxAttempts ( maxAttempts ) ; retry . setBackoffIntervalMillis ( backoffIntervalMillis ) ; retry . setFailTimeoutMillis ( Long . MAX_VALUE ) ; auditProperties . setRetry ( retry ) ; datawave . microservice . audit . AuditServiceTest . isHealthy = false ; java . util . Collection < java . lang . String > roles = java . util . Collections . singleton ( "AuthorizedUser" ) ; datawave . security . authorization . DatawaveUser uathDWUser = new datawave . security . authorization . DatawaveUser ( DN , USER , null , roles , null , java . lang . System . currentTimeMillis ( ) ) ; datawave . microservice . authorization . user . ProxiedUserDetails authUser = new datawave . microservice . authorization . user . ProxiedUserDetails ( java . util . Collections . singleton ( uathDWUser ) , uathDWUser . getCreationTime ( ) ) ; org . springframework . web . util . UriComponents uri = org . springframework . web . util . UriComponentsBuilder . newInstance ( ) . scheme ( "https" ) . host ( "localhost" ) . port ( webServicePort ) . path ( "/audit/v1/audit" ) . build ( ) ; org . springframework . util . MultiValueMap < java . lang . String , java . lang . String > map = new org . springframework . util . LinkedMultiValueMap ( ) ; map . add ( AuditParameters . USER_DN , userDN ) ; map . add ( AuditParameters . QUERY_STRING , query ) ; map . add ( AuditParameters . QUERY_AUTHORIZATIONS , authorizations ) ; map . add ( AuditParameters . QUERY_AUDIT_TYPE , auditType . name ( ) ) ; map . add ( AuditParameters . QUERY_SECURITY_MARKING_COLVIZ , "ALL" ) ; org . springframework . http . RequestEntity requestEntity = jwtRestTemplate . createRequestEntity ( authUser , map , null , HttpMethod . POST , uri ) ; long startTimeMillis = java . lang . System . currentTimeMillis ( ) ; jwtRestTemplate . exchange ( requestEntity , java . lang . String . class ) ; long stopTimeMillis = java . lang . System . currentTimeMillis ( ) ; "<AssertPlaceHolder>" ; } createRequestEntity ( datawave . microservice . authorization . user . ProxiedUserDetails , T , org . springframework . util . MultiValueMap , org . springframework . http . HttpMethod , org . springframework . web . util . UriComponents ) { java . lang . String token = jwtTokenHandler . createTokenFromUsers ( currentUser . getUsername ( ) , currentUser . getProxiedUsers ( ) ) ; org . springframework . http . HttpHeaders headers = new org . springframework . http . HttpHeaders ( ) ; headers . add ( HttpHeaders . AUTHORIZATION , ( "Bearer<sp>" + token ) ) ; if ( additionalHeaders != null ) { additionalHeaders . forEach ( headers :: put ) ; } return new org . springframework . http . RequestEntity ( body , headers , method , uri . toUri ( ) ) ; }
org . junit . Assert . assertTrue ( ( ( stopTimeMillis - startTimeMillis ) >= ( maxAttempts * backoffIntervalMillis ) ) )
testEquals_equalsWithSameValues ( ) { org . eclipse . rap . rwt . template . Position position1 = new org . eclipse . rap . rwt . template . Position ( 3.14F , 42 ) ; org . eclipse . rap . rwt . template . Position position2 = new org . eclipse . rap . rwt . template . Position ( 3.14F , 42 ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) return true ; if ( obj == null ) return false ; if ( ( getClass ( ) ) != ( obj . getClass ( ) ) ) return false ; org . eclipse . rap . e4 . internal . RAPEventObjectSupplier . Subscriber other = ( ( org . eclipse . rap . e4 . internal . RAPEventObjectSupplier . Subscriber ) ( obj ) ) ; if ( ( requestor ) == null ) { if ( ( other . requestor ) != null ) return false ; } else if ( ! ( requestor . equals ( other . requestor ) ) ) return false ; if ( ( topic ) == null ) { if ( ( other . topic ) != null ) return false ; } else if ( ! ( topic . equals ( other . topic ) ) ) return false ; return true ; }
org . junit . Assert . assertTrue ( position1 . equals ( position2 ) )
serialize ( ) { com . google . gson . Gson gson = com . github . seratch . jslack . common . json . GsonFactory . createSnakeCase ( ) ; com . github . seratch . jslack . api . model . event . GridMigrationStartedEvent event = new com . github . seratch . jslack . api . model . event . GridMigrationStartedEvent ( ) ; java . lang . String generatedJson = gson . toJson ( event ) ; java . lang . String expectedJson = "{\"type\":\"grid_migration_started\"}" ; "<AssertPlaceHolder>" ; } createSnakeCase ( ) { return new com . google . gson . GsonBuilder ( ) . setFieldNamingPolicy ( FieldNamingPolicy . LOWER_CASE_WITH_UNDERSCORES ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . LayoutBlock . class , new com . github . seratch . jslack . common . json . GsonLayoutBlockFactory ( ) ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . composition . TextObject . class , new com . github . seratch . jslack . common . json . GsonTextObjectFactory ( ) ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . ContextBlockElement . class , new com . github . seratch . jslack . common . json . GsonContextBlockElementFactory ( ) ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . element . BlockElement . class , new com . github . seratch . jslack . common . json . GsonBlockElementFactory ( ) ) . create ( ) ; }
org . junit . Assert . assertThat ( generatedJson , org . hamcrest . CoreMatchers . is ( expectedJson ) )
shouldGPS ( ) { java . lang . String regex = "([0-9]{3}\\.[0-9]{3}\\,[0-9]{3}\\.[0-9]{3})" ; for ( int i = 0 ; i < 100 ; i ++ ) { java . lang . String text = com . github . jsonzou . jmockdata . util . RandomUtils . nextStringFromRegex ( regex ) ; System . out . println ( text ) ; "<AssertPlaceHolder>" ; } } nextStringFromRegex ( java . lang . String ) { return com . github . jsonzou . jmockdata . util . RandomUtils . REGEX_GENERATOR . generateByRegex ( regex ) ; }
org . junit . Assert . assertTrue ( text . matches ( regex ) )
testJDBCSQLVTI ( ) { java . lang . String sql = java . lang . String . format ( ( "select<sp>*<sp>from<sp>table<sp>(JDBCSQLVTI('jdbc:splice://localhost:1527/splicedb;create=true;" + ( "user=splice;password=admin',<sp>" + "'select<sp>*<sp>from<sp>%s.%s'))a" ) ) , com . splicemachine . derby . impl . sql . execute . operations . VTIOperationIT . CLASS_NAME , com . splicemachine . derby . impl . sql . execute . operations . VTIOperationIT . TABLE_NAME ) ; java . sql . ResultSet rs = com . splicemachine . derby . impl . sql . execute . operations . VTIOperationIT . spliceClassWatcher . executeQuery ( sql ) ; int count = 0 ; while ( rs . next ( ) ) { count ++ ; } "<AssertPlaceHolder>" ; } next ( ) { return stepNext ( true ) ; }
org . junit . Assert . assertEquals ( 2 , count )
addMultiWordCrossFailTest ( ) { this . crossword . addWord ( new de . unima . dws . semanta . crossword . model . HAWord ( new de . unima . dws . semanta . model . HAEntity ( null , "test" , null , null , null ) ) , 0 , 1 , 3 , 1 ) ; this . crossword . addWord ( new de . unima . dws . semanta . crossword . model . HAWord ( new de . unima . dws . semanta . model . HAEntity ( null , "semanta" , null , null , null ) ) , 2 , 0 , 2 , 6 ) ; "<AssertPlaceHolder>" ; } size ( ) { return words . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , this . crossword . size ( ) )
testFindByPrimaryKeyExisting ( ) { com . liferay . asset . tag . stats . model . AssetTagStats newAssetTagStats = addAssetTagStats ( ) ; com . liferay . asset . tag . stats . model . AssetTagStats existingAssetTagStats = _persistence . findByPrimaryKey ( newAssetTagStats . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
org . junit . Assert . assertEquals ( existingAssetTagStats , newAssetTagStats )
testTaskRequestWithOldStartMsGetsUpdated ( ) { org . apache . kafka . common . utils . MockTime time = new org . apache . kafka . common . utils . MockTime ( 0 , 0 , 0 ) ; org . apache . kafka . common . utils . Scheduler scheduler = new org . apache . kafka . common . utils . MockScheduler ( time ) ; try ( org . apache . kafka . trogdor . common . MiniTrogdorCluster cluster = new org . apache . kafka . trogdor . common . MiniTrogdorCluster . Builder ( ) . addCoordinator ( "node01" ) . addAgent ( "node02" ) . scheduler ( scheduler ) . build ( ) ) { org . apache . kafka . trogdor . task . NoOpTaskSpec fooSpec = new org . apache . kafka . trogdor . task . NoOpTaskSpec ( 1 , 500 ) ; time . sleep ( 552 ) ; org . apache . kafka . trogdor . coordinator . CoordinatorClient coordinatorClient = cluster . coordinatorClient ( ) ; org . apache . kafka . trogdor . task . NoOpTaskSpec updatedSpec = new org . apache . kafka . trogdor . task . NoOpTaskSpec ( 552 , 500 ) ; coordinatorClient . createTask ( new org . apache . kafka . trogdor . rest . CreateTaskRequest ( "fooSpec" , fooSpec ) ) ; org . apache . kafka . trogdor . rest . TaskState expectedState = new org . apache . kafka . trogdor . common . ExpectedTasks . ExpectedTaskBuilder ( "fooSpec" ) . taskState ( new org . apache . kafka . trogdor . rest . TaskRunning ( updatedSpec , 552 , new com . fasterxml . jackson . databind . node . TextNode ( "receiving" ) ) ) . build ( ) . taskState ( ) ; org . apache . kafka . trogdor . rest . TaskState resp = coordinatorClient . task ( new org . apache . kafka . trogdor . rest . TaskRequest ( "fooSpec" ) ) ; "<AssertPlaceHolder>" ; } } task ( org . apache . kafka . trogdor . rest . TaskRequest ) { java . lang . String uri = javax . ws . rs . core . UriBuilder . fromPath ( url ( "/coordinator/tasks/{taskId}" ) ) . build ( request . taskId ( ) ) . toString ( ) ; org . apache . kafka . trogdor . rest . JsonRestServer . HttpResponse < org . apache . kafka . trogdor . rest . TaskState > resp = org . apache . kafka . trogdor . rest . JsonRestServer . httpRequest ( log , uri , "GET" , null , new com . fasterxml . jackson . core . type . TypeReference < org . apache . kafka . trogdor . rest . TaskState > ( ) { } , maxTries ) ; return resp . body ( ) ; }
org . junit . Assert . assertEquals ( expectedState , resp )
listenerExceptionAddAfterRunTest ( ) { org . threadly . test . concurrent . TestRunnable listener = new org . threadly . concurrent . TestRuntimeFailureRunnable ( ) ; org . threadly . concurrent . future . ListenableFutureTask < java . lang . Object > future = makeFutureTask ( org . threadly . concurrent . DoNothingRunnable . instance ( ) , null ) ; future . run ( ) ; try { future . listener ( listener ) ; org . junit . Assert . fail ( "Exception<sp>should<sp>have<sp>thrown" ) ; } catch ( java . lang . RuntimeException e ) { } "<AssertPlaceHolder>" ; } ranOnce ( ) { return ( runCount . get ( ) ) == 1 ; }
org . junit . Assert . assertTrue ( listener . ranOnce ( ) )
connectRepositoryAuthenticationInfo ( ) { this . wagon . connect ( this . repository , this . authenticationInfo ) ; "<AssertPlaceHolder>" ; verify ( this . sessionListenerSupport ) . fireSessionOpening ( ) ; verify ( this . wagon ) . connectToRepository ( this . repository , this . authenticationInfo , null ) ; verify ( this . sessionListenerSupport ) . fireSessionLoggedIn ( ) ; verify ( this . sessionListenerSupport ) . fireSessionOpened ( ) ; } getRepository ( ) { return this . repository ; }
org . junit . Assert . assertEquals ( this . repository , this . wagon . getRepository ( ) )
test_Issuer_Derived_Inequality ( ) { com . ripple . core . coretypes . Amount a = com . ripple . core . AmountTest . amt ( "1/USD/rNDKeo9RrCiRdfsMG8AdoZvNZxHASGzbZL" ) ; com . ripple . core . coretypes . Amount b = com . ripple . core . AmountTest . amt ( "1/USD/rH5aWQJ4R7v4Mpyf4kDBUvDFT5cbpFq3XP" ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ! ( o instanceof org . ripple . bouncycastle . asn1 . eac . PackedDate ) ) { return false ; } org . ripple . bouncycastle . asn1 . eac . PackedDate other = ( ( org . ripple . bouncycastle . asn1 . eac . PackedDate ) ( o ) ) ; return org . ripple . bouncycastle . util . Arrays . areEqual ( time , other . time ) ; }
org . junit . Assert . assertFalse ( a . equals ( b ) )
getAccessInfos ( ) { final java . lang . String id = "asyncSubscriptionId" ; subscribeAsync ( id ) ; java . lang . String orgId = tpAndSupplier . getOrganizationId ( ) ; container . login ( java . lang . String . valueOf ( supplierUser . getKey ( ) ) , org . oscm . subscriptionservice . dao . ROLE_TECHNOLOGY_MANAGER ) ; final org . oscm . internal . vo . VOInstanceInfo instanceInfo = new org . oscm . internal . vo . VOInstanceInfo ( ) ; instanceInfo . setInstanceId ( "completionProductInstanceId" ) ; instanceInfo . setAccessInfo ( "Public<sp>DNS<sp>for<sp>EC2<sp>instance:<sp>ec2-66-66-66-66.compute-1.amazonaws.com" ) ; subMgmt . completeAsyncSubscription ( id , orgId , instanceInfo ) ; java . util . List < java . lang . String > result = runTX ( new java . util . concurrent . Callable < java . util . List < java . lang . String > > ( ) { @ org . oscm . subscriptionservice . dao . Override public java . util . List < java . lang . String > call ( ) throws org . oscm . subscriptionservice . dao . Exception { org . oscm . domobjects . Subscription qryObj = new org . oscm . domobjects . Subscription ( ) ; qryObj . setOrganizationKey ( tpAndSupplier . getKey ( ) ) ; qryObj . setSubscriptionId ( id ) ; org . oscm . domobjects . Subscription subscription = ( ( org . oscm . domobjects . Subscription ) ( mgr . find ( qryObj ) ) ) ; return dao . getAccessInfos ( subscription , instanceInfo ) ; } } ) ; "<AssertPlaceHolder>" ; } size ( ) { return categoriesForMarketplace . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , result . size ( ) )
testValidateSignedZip ( ) { boolean result = zipValidator . validateZipFile ( org . codice . ddf . catalog . transformer . zip . ZipValidatorTest . SIGNED_ZIP_PATH ) ; "<AssertPlaceHolder>" ; } validateZipFile ( java . lang . String ) { try ( java . util . jar . JarFile jarFile = new java . util . jar . JarFile ( filePath ) ) { java . util . jar . Manifest man = jarFile . getManifest ( ) ; if ( man == null ) { throw new org . codice . ddf . catalog . transformer . zip . ZipValidationException ( "Zip<sp>validation<sp>failed,<sp>missing<sp>manifest<sp>file." ) ; } java . util . List < java . util . jar . JarEntry > entriesVec = new java . util . ArrayList ( ) ; byte [ ] buffer = new byte [ ZipDecompression . BUFFER_SIZE ] ; java . util . Enumeration entries = jarFile . entries ( ) ; while ( entries . hasMoreElements ( ) ) { java . util . jar . JarEntry je = ( ( java . util . jar . JarEntry ) ( entries . nextElement ( ) ) ) ; if ( je . isDirectory ( ) ) { continue ; } entriesVec . add ( je ) ; try ( java . io . InputStream is = jarFile . getInputStream ( je ) ) { while ( ( is . read ( buffer , 0 , buffer . length ) ) != ( - 1 ) ) { } is . close ( ) ; } catch ( java . io . IOException e ) { throw new org . codice . ddf . catalog . transformer . zip . ZipValidationException ( java . lang . String . format ( "Zip<sp>validation<sp>failed,<sp>unable<sp>to<sp>get<sp>input<sp>stream<sp>for<sp>entry<sp>%s" , je . getName ( ) ) ) ; } } java . util . Iterator < java . util . jar . JarEntry > iter = entriesVec . iterator ( ) ; while ( iter . hasNext ( ) ) { java . util . jar . JarEntry je = iter . next ( ) ; java . security . cert . Certificate [ ] certs = je . getCertificates ( ) ; if ( ( certs == null ) || ( ( certs . length ) == 0 ) ) { if ( ! ( je . getName ( ) . startsWith ( "META-INF" ) ) ) { throw new org . codice . ddf . catalog . transformer . zip . ZipValidationException ( java . lang . String . format ( "Zip<sp>validation<sp>failed,<sp>unable<sp>to<sp>get<sp>certificates<sp>for<sp>entry<sp>%s" , je . getName ( ) ) ) ; } } else { int startIndex = 0 ; java . security . cert . X509Certificate [ ] certChain ; while ( ( certChain = getAChain ( certs , startIndex ) ) != null ) { try { merlin . verifyTrust ( certChain [ 0 ] . getPublicKey ( ) ) ; } catch ( org . apache . wss4j . common . ext . WSSecurityException e1 ) { throw new org . codice . ddf . catalog . transformer . zip . ZipValidationException ( java . lang . String . format ( "Zip<sp>validation<sp>failed,<sp>untrusted<sp>certificates<sp>for<sp>entry<sp>%s" , je . getName ( ) ) ) ; } startIndex += certChain . length ; } } } } catch ( java . io . IOException e ) { throw new org . codice . ddf . catalog . transformer . zip . ZipValidationException ( java . lang . String . format ( "Zip<sp>validation<sp>failed<sp>for<sp>file<sp>:<sp>%s" , filePath ) ) ; } return true ; }
org . junit . Assert . assertThat ( result , org . hamcrest . CoreMatchers . is ( true ) )
sharedPaymentMethodToken ( ) { com . braintreegateway . integrationtest . BraintreeGateway sharerGateway = new com . braintreegateway . integrationtest . BraintreeGateway ( Environment . DEVELOPMENT , "integration_merchant_public_id" , "oauth_app_partner_user_public_key" , "oauth_app_partner_user_private_key" ) ; com . braintreegateway . integrationtest . Customer customer = sharerGateway . customer ( ) . create ( new com . braintreegateway . integrationtest . CustomerRequest ( ) . creditCard ( ) . number ( "5105105105105100" ) . expirationDate ( "05/19" ) . billingAddress ( ) . postalCode ( "94107" ) . done ( ) . done ( ) ) . getTarget ( ) ; com . braintreegateway . integrationtest . CreditCard card = customer . getCreditCards ( ) . get ( 0 ) ; com . braintreegateway . integrationtest . Address billingAddress = card . getBillingAddress ( ) ; com . braintreegateway . integrationtest . Address shippingAddress = sharerGateway . address ( ) . create ( customer . getId ( ) , new com . braintreegateway . integrationtest . AddressRequest ( ) . postalCode ( "94107" ) ) . getTarget ( ) ; com . braintreegateway . integrationtest . BraintreeGateway oauthGateway = new com . braintreegateway . integrationtest . BraintreeGateway ( "client_id$development$integration_client_id" , "client_secret$development$integration_client_secret" ) ; java . lang . String code = com . braintreegateway . testhelpers . TestHelper . createOAuthGrant ( oauthGateway , "integration_merchant_id" , "shared_vault_transactions" ) ; com . braintreegateway . integrationtest . OAuthCredentialsRequest oauthRequest = new com . braintreegateway . integrationtest . OAuthCredentialsRequest ( ) . code ( code ) . scope ( "shared_vault_transactions" ) ; com . braintreegateway . integrationtest . Result < com . braintreegateway . integrationtest . OAuthCredentials > accessTokenResult = oauthGateway . oauth ( ) . createTokenFromCode ( oauthRequest ) ; com . braintreegateway . integrationtest . BraintreeGateway gateway = new com . braintreegateway . integrationtest . BraintreeGateway ( accessTokenResult . getTarget ( ) . getAccessToken ( ) ) ; com . braintreegateway . integrationtest . TransactionRequest request = new com . braintreegateway . integrationtest . TransactionRequest ( ) . amount ( TransactionAmount . AUTHORIZE . amount ) . sharedPaymentMethodToken ( card . getToken ( ) ) . sharedCustomerId ( customer . getId ( ) ) . sharedShippingAddressId ( shippingAddress . getId ( ) ) . sharedBillingAddressId ( billingAddress . getId ( ) ) ; com . braintreegateway . integrationtest . Result < com . braintreegateway . integrationtest . Transaction > result = gateway . transaction ( ) . sale ( request ) ; "<AssertPlaceHolder>" ; } isSuccess ( ) { return success ; }
org . junit . Assert . assertTrue ( result . isSuccess ( ) )
testIsComplexType ( ) { "<AssertPlaceHolder>" ; } isComplexType ( ) { return true ; }
org . junit . Assert . assertFalse ( this . simple . isComplexType ( ) )
testToDBObject_Expression ( ) { final org . apache . rya . mongodb . document . visibility . DocumentVisibility dv = new org . apache . rya . mongodb . document . visibility . DocumentVisibility ( "A&B&C" ) ; final com . mongodb . BasicDBObject dbObject = org . apache . rya . mongodb . document . visibility . DocumentVisibilityAdapter . toDBObject ( dv . getExpression ( ) ) ; final com . mongodb . BasicDBObject expected = ( ( com . mongodb . BasicDBObject ) ( com . mongodb . util . JSON . parse ( ( "{" + ( "documentVisibility<sp>:<sp>[[\"A\",<sp>\"B\",<sp>\"C\"]]" + "}" ) ) ) ) ) ; "<AssertPlaceHolder>" ; } parse ( java . lang . String ) { checkNotNull ( requestText ) ; int startOfSparql = 0 ; final java . util . Set < org . apache . rya . indexing . pcj . storage . accumulo . VariableOrder > varOrders = new java . util . HashSet ( ) ; final java . util . regex . Matcher matcher = org . apache . rya . indexing . pcj . fluo . client . util . ParsedQueryRequest . varOrdersPattern . matcher ( requestText ) ; while ( matcher . find ( ) ) { final java . lang . String varOrder = matcher . group ( 1 ) ; varOrders . add ( new org . apache . rya . indexing . pcj . storage . accumulo . VariableOrder ( varOrder . split ( ",\\s*" ) ) ) ; startOfSparql = matcher . end ( ) ; } final java . lang . String sparql = requestText . substring ( startOfSparql ) . trim ( ) ; return new org . apache . rya . indexing . pcj . fluo . client . util . ParsedQueryRequest ( sparql , varOrders ) ; }
org . junit . Assert . assertEquals ( expected , dbObject )
testExpireAt ( ) { org . redisson . api . RSetCache < java . lang . String > cache = redisson . getSetCache ( "simple" ) ; cache . add ( "8" , 1 , TimeUnit . SECONDS ) ; cache . expireAt ( ( ( java . lang . System . currentTimeMillis ( ) ) + 100 ) ) ; java . lang . Thread . sleep ( 500 ) ; "<AssertPlaceHolder>" ; cache . destroy ( ) ; } size ( ) { return map . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , cache . size ( ) )
testExpressionLikeLangauge ( ) { org . eclipse . xtend2 . lib . StringConcatenation _builder = new org . eclipse . xtend2 . lib . StringConcatenation ( ) ; _builder . append ( "grammar<sp>test<sp>with<sp>org.eclipse.xtext.common.Terminals<sp>generate<sp>test<sp>\'http://test\'" ) ; java . lang . String grammar = _builder . toString ( ) ; java . lang . String _grammar = grammar ; org . eclipse . xtend2 . lib . StringConcatenation _builder_1 = new org . eclipse . xtend2 . lib . StringConcatenation ( ) ; _builder_1 . append ( "<sp>" ) ; _builder_1 . append ( "Ex<sp>:\tAtom<sp>({ChainExpression.left+=current}<sp>operator=(\'+\'|\'-\'|\'*\'|\'/\')<sp>right=Atom<sp>)*;Atom<sp>returns<sp>Ex<sp>:<sp>Number<sp>|<sp>\'(\'<sp>Ex<sp>\')\';Number<sp>:<sp>value=INT;" ) ; grammar = _grammar + _builder_1 ; org . eclipse . emf . ecore . EPackage ePackage = this . getEPackageFromGrammar ( grammar ) ; org . eclipse . emf . ecore . EClassifier _type = this . < org . eclipse . emf . ecore . EClassifier > type ( ePackage , "Ex" ) ; org . eclipse . emf . ecore . EClass classifier = ( ( org . eclipse . emf . ecore . EClass ) ( _type ) ) ; "<AssertPlaceHolder>" ; } features ( org . eclipse . emf . ecore . EStructuralFeature [ ] ) { java . util . Set < com . google . common . base . Predicate < org . eclipse . xtext . formatting2 . regionaccess . ISemanticRegion > > predicates = com . google . common . collect . Sets . newHashSet ( ) ; for ( int i = 0 ; i < ( features . length ) ; i ++ ) { org . eclipse . emf . ecore . EStructuralFeature feat = features [ i ] ; assertNoContainment ( feat ) ; predicates . add ( new org . eclipse . xtext . formatting2 . regionaccess . internal . AbstractSemanticRegionsFinder . FeaturePredicate ( feat ) ) ; } return findAll ( com . google . common . base . Predicates . or ( predicates ) ) ; }
org . junit . Assert . assertEquals ( 0 , this . features ( classifier ) . size ( ) )
qlAddressBeginswithAndSexFulltext ( ) { java . lang . String sqlite = "SELECT<sp>*<sp>FROM<sp>users<sp>WHERE<sp>(address<sp>LIKE<sp>'D%'<sp>and<sp>sex<sp>LIKE<sp>'%male%')<sp>LIMIT<sp>10" ; java . lang . String api = "select<sp>*<sp>where<sp>address<sp>=<sp>'D*'<sp>and<sp>sex<sp>contains<sp>'male'" ; org . apache . usergrid . query . validator . QueryRequest request = new org . apache . usergrid . query . validator . QueryRequest ( ) ; request . setDbQuery ( sqlite ) ; request . getApiQuery ( ) . setQuery ( api ) ; org . apache . usergrid . query . validator . QueryResponse response = validator . execute ( request , new org . apache . usergrid . query . validator . QueryResultsMatcher ( ) { @ org . apache . usergrid . query . validator . users . Override public boolean equals ( java . util . List < org . apache . usergrid . persistence . Entity > expectedEntities , java . util . List < org . apache . usergrid . persistence . Entity > actuallyEntities ) { boolean equals = ( expectedEntities . size ( ) ) == ( expectedEntities . size ( ) ) ; if ( ! equals ) return false ; for ( org . apache . usergrid . persistence . Entity entity : actuallyEntities ) { java . lang . String address = ( ( java . lang . String ) ( entity . getProperty ( "address" ) ) ) ; if ( ( org . apache . usergrid . utils . StringUtils . startsWith ( address , "D" ) ) == false ) { return false ; } } return equals ; } } ) ; "<AssertPlaceHolder>" ; } toString ( ) { java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; if ( ( serviceName ) != null ) { sb . append ( "/" ) ; sb . append ( serviceName ) ; } for ( int i = 0 ; i < ( parameters . size ( ) ) ; i ++ ) { org . apache . usergrid . services . ServiceParameter p = parameters . get ( i ) ; if ( p instanceof org . apache . usergrid . services . ServiceParameter . QueryParameter ) { if ( i == ( ( parameters . size ( ) ) - 1 ) ) { sb . append ( '?' ) ; } else { sb . append ( ';' ) ; } boolean has_prev_param = false ; java . lang . String q = p . toString ( ) ; if ( isNotBlank ( q ) ) { try { sb . append ( "ql=" ) . append ( java . net . URLEncoder . encode ( q , "UTF-8" ) ) ; } catch ( java . io . UnsupportedEncodingException e ) { org . apache . usergrid . services . ServiceRequest . logger . error ( "Unable<sp>to<sp>encode<sp>url" , e ) ; } has_prev_param = true ; } int limit = p . getQuery ( ) . getLimit ( ) ; if ( limit != ( org . apache . usergrid . persistence . Query . DEFAULT_LIMIT ) ) { if ( has_prev_param ) { sb . append ( '&' ) ; } sb . append ( "limit=" ) . append ( limit ) ; has_prev_param = true ; } if ( ( p . getQuery ( ) . getStartResult ( ) ) != null ) { if ( has_prev_param ) { sb . append ( '&' ) ; } sb . append ( "start=" ) . append ( p . getQuery ( ) . getStartResult ( ) ) ; has_prev_param = true ; } } else { sb . append ( '/' ) ; sb . append ( p . toString ( ) ) ; } } return sb . toString ( ) ; }
org . junit . Assert . assertTrue ( response . toString ( ) , response . result ( ) )
templateTrailing$ ( ) { java . lang . String orig = ( com . sun . btrace . ArgsMapTest . KEY1 ) + "$" ; java . lang . String value = instance . template ( orig ) ; "<AssertPlaceHolder>" ; } template ( java . lang . String ) { if ( value == null ) { return null ; } if ( value . isEmpty ( ) ) { return value ; } java . util . regex . Matcher matcher = com . sun . btrace . ArgsMap . PatternSingleton . INSTANCE . matcher ( value ) ; java . lang . StringBuffer buffer = new java . lang . StringBuffer ( value . length ( ) ) ; while ( matcher . find ( ) ) { java . lang . String val = get ( matcher . group ( 1 ) ) ; matcher . appendReplacement ( buffer , ( val != null ? val : "$0" ) ) ; } matcher . appendTail ( buffer ) ; return buffer . toString ( ) ; }
org . junit . Assert . assertEquals ( orig , value )
cssMerge__when_merging_to_NULL_OBJECT_should_return_NULL_OBJECT2 ( ) { io . github . seleniumquery . by . secondgen . finder . CssFinder mergedCss = CssFinder . CSS_NOT_NATIVELY_SUPPORTED . merge ( io . github . seleniumquery . by . secondgen . finder . CssFinder . universalSelector ( ) ) ; "<AssertPlaceHolder>" ; } is ( java . lang . String ) { return isAnd ( IsEvaluator . IS_EVALUATOR , selector ) ; }
org . junit . Assert . assertThat ( mergedCss , org . hamcrest . CoreMatchers . is ( CssFinder . CSS_NOT_NATIVELY_SUPPORTED ) )
defaultTest2 ( ) { com . zaxxer . hikari . HikariDataSource dataSource = new com . zaxxer . hikari . HikariDataSource ( ) ; dataSource . setDataSourceClassName ( com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . DATA_SOURCE_CLASS_NAME ) ; dataSource . addDataSourceProperty ( "url" , com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . JDBC_URL ) ; try { java . sql . Connection connection = dataSource . getConnection ( ) ; "<AssertPlaceHolder>" ; java . lang . Thread . sleep ( 500 ) ; connection . close ( ) ; java . lang . Thread . sleep ( 500 ) ; com . navercorp . pinpoint . bootstrap . plugin . test . PluginTestVerifier verifier = com . navercorp . pinpoint . bootstrap . plugin . test . PluginTestVerifierHolder . getInstance ( ) ; verifier . printCache ( ) ; verifier . verifyTrace ( event ( com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . serviceType , "com.zaxxer.hikari.HikariDataSource.HikariDataSource()" ) ) ; verifier . verifyTrace ( event ( com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . serviceType , com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . getConnectionMethod1 ) ) ; verifier . verifyTrace ( event ( com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . serviceType , "com.zaxxer.hikari.pool.BaseHikariPool.BaseHikariPool(com.zaxxer.hikari.HikariConfig,<sp>java.lang.String,<sp>java.lang.String)" ) ) ; verifier . verifyTrace ( event ( com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . serviceType , com . navercorp . pinpoint . plugin . hikaricp . HikariCpIT . proxyConnectionMethod ) ) ; } finally { if ( dataSource != null ) { dataSource . close ( ) ; } } } getConnection ( ) { throw new java . lang . UnsupportedOperationException ( ) ; }
org . junit . Assert . assertNotNull ( connection )
testExpm1Definition ( ) { double epsilon = 3.0E-16 ; for ( int maxOrder = 0 ; maxOrder < 5 ; ++ maxOrder ) { org . hipparchus . analysis . differentiation . DSFactory factory = new org . hipparchus . analysis . differentiation . DSFactory ( 1 , maxOrder ) ; for ( double x = 0.1 ; x < 1.2 ; x += 0.001 ) { org . hipparchus . analysis . differentiation . DerivativeStructure dsX = factory . variable ( 0 , x ) ; org . hipparchus . analysis . differentiation . DerivativeStructure expm11 = org . hipparchus . util . FastMath . expm1 ( dsX ) ; org . hipparchus . analysis . differentiation . DerivativeStructure expm12 = dsX . exp ( ) . subtract ( dsX . getField ( ) . getOne ( ) ) ; org . hipparchus . analysis . differentiation . DerivativeStructure zero = expm11 . subtract ( expm12 ) ; for ( int n = 0 ; n <= maxOrder ; ++ n ) { "<AssertPlaceHolder>" ; } } } } getPartialDerivative ( int [ ] ) { return data [ getFactory ( ) . getCompiler ( ) . getPartialDerivativeIndex ( orders ) ] ; }
org . junit . Assert . assertEquals ( 0 , zero . getPartialDerivative ( n ) , epsilon )
createBlackTest ( ) { final net . violet . platform . datamodel . User mainUser = Factories . USER . find ( 1 ) ; final net . violet . platform . datamodel . User blackUser = Factories . USER . find ( 2 ) ; Factories . BLACK . createNewBlack ( mainUser , blackUser ) ; "<AssertPlaceHolder>" ; } getBlackList ( ) { if ( ( this . uBlackList ) == null ) { try { this . uBlackList = net . violet . db . records . associations . DecoratedAssociation . createDecoratedAssociation ( this , net . violet . platform . datamodel . UserImpl . SPECIFICATION , "black_blacked" , BlackImpl . SPECIFICATION , "black_user" ) ; } catch ( final java . sql . SQLException e ) { net . violet . platform . datamodel . UserImpl . LOGGER . fatal ( e , e ) ; } } return this . uBlackList ; }
org . junit . Assert . assertTrue ( mainUser . getBlackList ( ) . containsKey ( blackUser ) )
testGetP2RepURIFromCompFile_other ( ) { java . io . File compFile = java . io . File . createTempFile ( "testcomp.text" , null ) ; java . net . URI p2RepURI = org . talend . updates . runtime . utils . PathUtils . getP2RepURIFromCompFile ( compFile ) ; "<AssertPlaceHolder>" ; compFile . delete ( ) ; } getP2RepURIFromCompFile ( java . io . File ) { if ( compFile == null ) { return null ; } final java . lang . String name = compFile . getName ( ) . toLowerCase ( ) ; if ( ( ( name . endsWith ( FileExtensions . JAR_FILE_SUFFIX ) ) || ( name . endsWith ( FileExtensions . ZIP_FILE_SUFFIX ) ) ) || ( name . endsWith ( FileExtensions . CAR_FILE_SUFFIX ) ) ) { return java . net . URI . create ( ( ( "jar:" + ( compFile . toURI ( ) . toString ( ) ) ) + "!/" ) ) ; } return null ; }
org . junit . Assert . assertNull ( p2RepURI )
serializeBroadcastSubscriptionRequest ( ) { java . lang . String subscriptionId = "1234" ; java . lang . String subscribedToName = "myEvent" ; testBroadcastInterface . LocationUpdateWithSpeedSelectiveBroadcastFilterParameters filterParameters = new testBroadcastInterface . LocationUpdateWithSpeedSelectiveBroadcastFilterParameters ( ) ; filterParameters . setCountry ( "Germany" ) ; filterParameters . setStartTime ( "4:00" ) ; joynr . OnChangeSubscriptionQos qos = new joynr . OnChangeSubscriptionQos ( ) . setMinIntervalMs ( 0 ) . setValidityMs ( 60000 ) . setPublicationTtlMs ( 1000 ) ; joynr . BroadcastSubscriptionRequest broadcastSubscription = new joynr . BroadcastSubscriptionRequest ( subscriptionId , subscribedToName , filterParameters , qos ) ; java . lang . String writeValueAsString = objectMapper . writeValueAsString ( broadcastSubscription ) ; System . out . println ( writeValueAsString ) ; joynr . BroadcastSubscriptionRequest receivedbroadcastSubscription = objectMapper . readValue ( writeValueAsString , joynr . BroadcastSubscriptionRequest . class ) ; "<AssertPlaceHolder>" ; } setPublicationTtlMs ( long ) { return ( ( joynr . PeriodicSubscriptionQos ) ( super . setPublicationTtlMs ( publicationTtlMs ) ) ) ; }
org . junit . Assert . assertEquals ( broadcastSubscription , receivedbroadcastSubscription )
resolveArtifactWithLocalRepoFromSettings ( ) { java . lang . String localRepoPath = "target/local-repo-settings" ; java . lang . System . setProperty ( "org.ops4j.pax.url.mvn.settings" , "src/test/resources/settings-local-repo.xml" ) ; java . io . File localRepo = new java . io . File ( localRepoPath ) ; localRepo . mkdirs ( ) ; java . net . URL url = new java . net . URL ( "mvn:org.ops4j.base/ops4j-base-lang/1.0.0" ) ; url . openStream ( ) . close ( ) ; java . io . File artifact = new java . io . File ( localRepo , "org/ops4j/base/ops4j-base-lang/1.0.0/ops4j-base-lang-1.0.0.jar" ) ; "<AssertPlaceHolder>" ; } close ( ) { m_client . close ( ) ; }
org . junit . Assert . assertThat ( artifact . exists ( ) , org . hamcrest . CoreMatchers . is ( true ) )
findSmileysShouldFindTextWithOnlyASmiley ( ) { final java . util . List < net . usikkert . kouchat . android . smiley . Smiley > smileys = locator . findSmileys ( ":)" ) ; "<AssertPlaceHolder>" ; verifySmiley ( smileys , ":)" , 0 , 2 ) ; } size ( ) { return userList . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , smileys . size ( ) )
escapeString ( ) { final java . lang . String [ ] testPairs = new java . lang . String [ ] { "" , "" , "plain<sp>string" , "plain<sp>string" , "string<sp>with<sp>\"quotes\"" , "string<sp>with<sp>\"quotes\"" , "string<sp>with<sp>'quotes'" , "string<sp>with<sp>'quotes'" , "string<sp>with<sp>'quotes'" , "string<sp>with<sp>'quotes'" , "C:\\Program<sp>Files\\MyProgram" , "C:\\\\Program<sp>Files\\\\MyProgram" , "string\nwith\nnewlines" , "string\\nwith\\nnewlines" , "string\twith\ttabs" , "string\\twith\\ttabs" } ; for ( int i = 0 ; i < ( testPairs . length ) ; i += 2 ) { "<AssertPlaceHolder>" ; } } escapeString ( java . lang . String ) { final java . lang . StringBuilder result = new java . lang . StringBuilder ( ( 3 * ( str . length ( ) ) ) ) ; for ( int i = 0 ; i < ( str . length ( ) ) ; i ++ ) { char c = str . charAt ( i ) ; if ( c < ( com . google . gerrit . server . ioutil . StringUtil . NON_PRINTABLE_CHARS . length ) ) { result . append ( com . google . gerrit . server . ioutil . StringUtil . NON_PRINTABLE_CHARS [ c ] ) ; } else if ( c == '\\' ) { result . append ( "\\\\" ) ; } else { result . append ( c ) ; } } return result . toString ( ) ; }
org . junit . Assert . assertEquals ( com . google . gerrit . server . ioutil . StringUtil . escapeString ( testPairs [ i ] ) , testPairs [ ( i + 1 ) ] )
testIfftshiftComplex_inplace ( ) { double [ ] shiftVectorTemp_EXPECTED = org . jlinda . core . utils . SpectralUtilsTest . shiftVector_EXPECTED . clone ( ) ; org . jblas . ComplexDoubleMatrix matrixCplx_ACTUAL = new org . jblas . ComplexDoubleMatrix ( new org . jblas . DoubleMatrix ( shiftVectorTemp_EXPECTED ) , new org . jblas . DoubleMatrix ( shiftVectorTemp_EXPECTED ) ) ; org . jblas . ComplexDoubleMatrix matrixCplx_EXPECTED = new org . jblas . ComplexDoubleMatrix ( new org . jblas . DoubleMatrix ( org . jlinda . core . utils . SpectralUtilsTest . vector_EXPECTED ) , new org . jblas . DoubleMatrix ( org . jlinda . core . utils . SpectralUtilsTest . vector_EXPECTED ) ) ; org . jlinda . core . utils . SpectralUtils . ifftshift_inplace ( matrixCplx_ACTUAL ) ; "<AssertPlaceHolder>" ; } ifftshift_inplace ( org . jblas . ComplexDoubleMatrix ) { inMatrix . copy ( org . jlinda . core . utils . SpectralUtils . ifftshift ( inMatrix ) ) ; }
org . junit . Assert . assertEquals ( matrixCplx_EXPECTED , matrixCplx_ACTUAL )
testEncryptDecrypt ( ) { com . google . crypto . tink . Aead aead = createInstance ( com . google . crypto . tink . subtle . Random . randBytes ( com . google . crypto . tink . subtle . XChaCha20Poly1305Test . KEY_SIZE ) ) ; for ( int i = 0 ; i < 100 ; i ++ ) { byte [ ] message = com . google . crypto . tink . subtle . Random . randBytes ( i ) ; byte [ ] aad = com . google . crypto . tink . subtle . Random . randBytes ( i ) ; byte [ ] ciphertext = aead . encrypt ( message , aad ) ; byte [ ] decrypted = aead . decrypt ( ciphertext , aad ) ; "<AssertPlaceHolder>" ; } } decrypt ( byte [ ] , byte [ ] ) { try { java . nio . ByteBuffer buffer = java . nio . ByteBuffer . wrap ( ciphertext ) ; int encryptedDekSize = buffer . getInt ( ) ; if ( ( encryptedDekSize <= 0 ) || ( encryptedDekSize > ( ( ciphertext . length ) - ( com . google . crypto . tink . aead . KmsEnvelopeAead . LENGTH_ENCRYPTED_DEK ) ) ) ) { throw new java . security . GeneralSecurityException ( "invalid<sp>ciphertext" ) ; } byte [ ] encryptedDek = new byte [ encryptedDekSize ] ; buffer . get ( encryptedDek , 0 , encryptedDekSize ) ; byte [ ] payload = new byte [ buffer . remaining ( ) ] ; buffer . get ( payload , 0 , buffer . remaining ( ) ) ; byte [ ] dek = remote . decrypt ( encryptedDek , com . google . crypto . tink . aead . KmsEnvelopeAead . EMPTY_AAD ) ; com . google . crypto . tink . Aead aead = com . google . crypto . tink . Registry . getPrimitive ( dekTemplate . getTypeUrl ( ) , dek , com . google . crypto . tink . Aead . class ) ; return aead . decrypt ( payload , associatedData ) ; } catch ( java . lang . IndexOutOfBoundsException | java . nio . BufferUnderflowException | java . lang . NegativeArraySizeException e ) { throw new java . security . GeneralSecurityException ( "invalid<sp>ciphertext" , e ) ; } }
org . junit . Assert . assertArrayEquals ( message , decrypted )
pluralFormOne ( ) { final org . apache . isis . applib . services . i18n . TranslatableString ts = org . apache . isis . applib . services . i18n . TranslatableString . trn ( "You<sp>can't<sp>do<sp>that<sp>because<sp>there<sp>is<sp>a<sp>dependent<sp>object" , "You<sp>can't<sp>do<sp>that<sp>because<sp>there<sp>are<sp>dependent<sp>objects" , 1 ) ; "<AssertPlaceHolder>" ; } getPattern ( ) { return pattern ; }
org . junit . Assert . assertThat ( ts . getPattern ( ) , org . hamcrest . CoreMatchers . is ( "You<sp>can't<sp>do<sp>that<sp>because<sp>there<sp>is<sp>a<sp>dependent<sp>object" ) )
testEncryptDecrypt ( ) { com . google . crypto . tink . Aead aead = getAead ( com . google . crypto . tink . subtle . Random . randBytes ( 16 ) , com . google . crypto . tink . subtle . Random . randBytes ( 16 ) , 16 , 16 , "HMACSHA256" ) ; byte [ ] plaintext = com . google . crypto . tink . subtle . Random . randBytes ( 1001 ) ; byte [ ] aad = com . google . crypto . tink . subtle . Random . randBytes ( 13 ) ; byte [ ] ciphertext = aead . encrypt ( plaintext , aad ) ; try { byte [ ] plaintext1 = aead . decrypt ( ciphertext , aad ) ; "<AssertPlaceHolder>" ; } catch ( java . security . GeneralSecurityException e ) { org . junit . Assert . fail ( ( "Valid<sp>ciphertext<sp>and<sp>aad,<sp>should<sp>have<sp>passed:<sp>" + e ) ) ; } } decrypt ( byte [ ] , byte [ ] ) { try { java . nio . ByteBuffer buffer = java . nio . ByteBuffer . wrap ( ciphertext ) ; int encryptedDekSize = buffer . getInt ( ) ; if ( ( encryptedDekSize <= 0 ) || ( encryptedDekSize > ( ( ciphertext . length ) - ( com . google . crypto . tink . aead . KmsEnvelopeAead . LENGTH_ENCRYPTED_DEK ) ) ) ) { throw new java . security . GeneralSecurityException ( "invalid<sp>ciphertext" ) ; } byte [ ] encryptedDek = new byte [ encryptedDekSize ] ; buffer . get ( encryptedDek , 0 , encryptedDekSize ) ; byte [ ] payload = new byte [ buffer . remaining ( ) ] ; buffer . get ( payload , 0 , buffer . remaining ( ) ) ; byte [ ] dek = remote . decrypt ( encryptedDek , com . google . crypto . tink . aead . KmsEnvelopeAead . EMPTY_AAD ) ; com . google . crypto . tink . Aead aead = com . google . crypto . tink . Registry . getPrimitive ( dekTemplate . getTypeUrl ( ) , dek , com . google . crypto . tink . Aead . class ) ; return aead . decrypt ( payload , associatedData ) ; } catch ( java . lang . IndexOutOfBoundsException | java . nio . BufferUnderflowException | java . lang . NegativeArraySizeException e ) { throw new java . security . GeneralSecurityException ( "invalid<sp>ciphertext" , e ) ; } }
org . junit . Assert . assertArrayEquals ( plaintext , plaintext1 )
shouldSumStrengthsOfConstituents ( ) { org . pitest . highwheel . model . Dependency testee = new org . pitest . highwheel . model . Dependency ( ) ; org . pitest . highwheel . model . AccessPoint source = org . pitest . highwheel . model . AccessPoint . create ( org . pitest . highwheel . model . ElementName . fromString ( "foo" ) ) ; org . pitest . highwheel . model . AccessPoint dest = org . pitest . highwheel . model . AccessPoint . create ( org . pitest . highwheel . model . ElementName . fromString ( "foo" ) ) ; testee . addDependency ( source , dest , AccessType . COMPOSED ) ; testee . addDependency ( source , dest , AccessType . USES ) ; "<AssertPlaceHolder>" ; } getStrength ( ) { return this . strength ; }
org . junit . Assert . assertEquals ( ( ( AccessType . COMPOSED . getStrength ( ) ) + ( AccessType . USES . getStrength ( ) ) ) , testee . getStrength ( ) )
testJoinRemovedAsParent ( ) { final J instance = getJoiningInstance ( "instance" ) ; final org . apache . oozie . fluentjob . api . dag . NodeBase child = new org . apache . oozie . fluentjob . api . dag . ExplicitNode ( "child" , null ) ; child . addParent ( instance ) ; child . removeParent ( instance ) ; "<AssertPlaceHolder>" ; } getChild ( ) { return child ; }
org . junit . Assert . assertEquals ( null , instance . getChild ( ) )
testGetBandDataType_uint8 ( ) { product . addBand ( "unsigned" , ProductData . TYPE_UINT8 ) ; final org . esa . beam . dataio . bigtiff . internal . TiffIFD tiffIFD = new org . esa . beam . dataio . bigtiff . internal . TiffIFD ( product ) ; "<AssertPlaceHolder>" ; } getBandDataType ( ) { return maxElemSizeBandDataType ; }
org . junit . Assert . assertEquals ( ProductData . TYPE_UINT8 , tiffIFD . getBandDataType ( ) )
whenPerformBinomialOnTwoIntegerValues_shouldReturnResultIfUnderInt ( ) { int result = com . google . common . math . IntMath . binomial ( 6 , 3 ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( 20 , result )
testGetDocument ( ) { cache . add ( list ) ; "<AssertPlaceHolder>" ; } getDocument ( ) { com . findwise . hydra . local . LocalDocument doc = new com . findwise . hydra . local . LocalDocument ( ) ; doc . putContentField ( "in1" , "in1" ) ; doc . putContentField ( "in2" , "in2" ) ; doc . putContentField ( "in3" , "in3" ) ; return doc ; }
org . junit . Assert . assertNotNull ( cache . getDocument ( ) )
testNamedKieBase ( ) { org . kie . api . KieBase kieBase = org . kie . spring . tests . KieSpringImportKieTest . context . getBean ( "kbase1" , org . kie . api . KieBase . class ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( kieBase )
testUnionPlan ( ) { org . apache . tajo . storage . fragment . FileFragment [ ] frags = org . apache . tajo . engine . planner . physical . StorageManager . splitNG ( org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . conf , "employee" , org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . employee . getMeta ( ) , org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . employee . getPath ( ) , Integer . MAX_VALUE ) ; org . apache . hadoop . fs . Path workDir = org . apache . tajo . util . CommonTestingUtil . getTestDir ( "target/test-data/testUnionPlan" ) ; org . apache . tajo . worker . TaskAttemptContext ctx = new org . apache . tajo . worker . TaskAttemptContext ( org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . conf , org . apache . tajo . LocalTajoTestingUtility . newQueryUnitAttemptId ( org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . masterPlan ) , new org . apache . tajo . storage . fragment . FileFragment [ ] { frags [ 0 ] } , workDir ) ; ctx . setEnforcer ( new org . apache . tajo . engine . planner . enforce . Enforcer ( ) ) ; org . apache . tajo . algebra . Expr context = org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . analyzer . parse ( QUERIES [ 0 ] ) ; org . apache . tajo . engine . planner . physical . LogicalPlan plan = org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . planner . createPlan ( context ) ; org . apache . tajo . engine . planner . physical . LogicalNode rootNode = org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . optimizer . optimize ( plan ) ; org . apache . tajo . engine . planner . physical . LogicalRootNode root = ( ( org . apache . tajo . engine . planner . physical . LogicalRootNode ) ( rootNode ) ) ; org . apache . tajo . engine . planner . physical . UnionNode union = plan . createNode ( org . apache . tajo . engine . planner . physical . UnionNode . class ) ; union . setLeftChild ( root . getChild ( ) ) ; union . setRightChild ( root . getChild ( ) ) ; root . setChild ( union ) ; org . apache . tajo . engine . planner . physical . PhysicalPlanner phyPlanner = new org . apache . tajo . engine . planner . physical . PhysicalPlannerImpl ( org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . conf , org . apache . tajo . engine . planner . physical . TestPhysicalPlanner . sm ) ; org . apache . tajo . engine . planner . physical . PhysicalExec exec = phyPlanner . createPlan ( ctx , root ) ; int count = 0 ; exec . init ( ) ; while ( ( exec . next ( ) ) != null ) { count ++ ; } exec . close ( ) ; "<AssertPlaceHolder>" ; } close ( ) { if ( ( resultSet ) != null ) { resultSet . close ( ) ; } resultSet = null ; isClosed = true ; }
org . junit . Assert . assertEquals ( 200 , count )
playerQualifiesWithNoUserIDReturnsFalse ( ) { when ( linker . getUserID ( player ) ) . thenReturn ( "" ) ; "<AssertPlaceHolder>" ; } playerQualifies ( org . bukkit . entity . Player , org . communitybridge . achievement . PlayerAchievementState ) { return false ; }
org . junit . Assert . assertFalse ( achievement . playerQualifies ( player , state ) )
testParseCookieMissingValue1 ( ) { final org . apache . hc . client5 . http . impl . cookie . RFC6265CookieSpec cookiespec = new org . apache . hc . client5 . http . impl . cookie . RFC6265CookieSpec ( ) ; final org . apache . hc . core5 . http . Header header = new org . apache . hc . core5 . http . message . BasicHeader ( "Set-Cookie" , "blah" ) ; final org . apache . hc . client5 . http . cookie . CookieOrigin origin = new org . apache . hc . client5 . http . cookie . CookieOrigin ( "host" , 80 , "/path/" , true ) ; final java . util . List < org . apache . hc . client5 . http . cookie . Cookie > cookies = cookiespec . parse ( header , origin ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . all . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , cookies . size ( ) )
testMember ( ) { org . gitlab4j . api . models . Member member = org . gitlab4j . api . JsonUtils . unmarshalResource ( org . gitlab4j . api . models . Member . class , "member.json" ) ; "<AssertPlaceHolder>" ; } compareJson ( T , java . lang . String ) { java . io . InputStreamReader reader = new java . io . InputStreamReader ( org . gitlab4j . api . TestGitLabApiBeans . class . getResourceAsStream ( filename ) ) ; return org . gitlab4j . api . JsonUtils . compareJson ( apiObject , reader ) ; }
org . junit . Assert . assertTrue ( org . gitlab4j . api . JsonUtils . compareJson ( member , "member.json" ) )
testArchiveZipLive ( ) { underTest . setArchive ( true ) ; underTest . setArxStreamFactory ( new org . dataconservancy . packaging . tool . impl . ZipArchiveStreamFactory ( ) ) ; java . io . ByteArrayOutputStream result = new java . io . ByteArrayOutputStream ( ) ; underTest . serialize ( state , StreamId . APPLICATION_VERSION , result ) ; "<AssertPlaceHolder>" ; } size ( ) { return size ; }
org . junit . Assert . assertTrue ( ( ( result . size ( ) ) > 1 ) )
testDeleteTenant ( ) { org . thingsboard . server . common . data . Tenant tenant = new org . thingsboard . server . common . data . Tenant ( ) ; tenant . setTitle ( "My<sp>tenant" ) ; org . thingsboard . server . common . data . Tenant savedTenant = tenantService . saveTenant ( tenant ) ; tenantService . deleteTenant ( savedTenant . getId ( ) ) ; org . thingsboard . server . common . data . Tenant foundTenant = tenantService . findTenantById ( savedTenant . getId ( ) ) ; "<AssertPlaceHolder>" ; } getId ( ) { return id ; }
org . junit . Assert . assertNull ( foundTenant )
testGetDatasourceFullWithLargeInterval ( ) { java . util . Map < java . lang . String , java . lang . Object > actual = resource . getDatasource ( dataSource , "m3" 7 , "m3" 3 ) ; java . util . Map < java . lang . String , java . lang . Object > expected = com . google . common . collect . ImmutableMap . < java . lang . String , java . lang . Object > builder ( ) . put ( "2014-02-13T00:00:00.000Z/2014-02-15T00:00:00.000Z" , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . put ( "m3" 2 , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . put ( "m3" 0 , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d2" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m2" ) ) ) . put ( "m3" 6 , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . put ( "2015-02-03T00:00:00.000Z/2015-02-05T00:00:00.000Z" , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" , "d2" , "m3" 5 ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" , "m2" , "m3" ) ) ) . put ( "2015-02-05T00:00:00.000Z/2015-02-09T00:00:00.000Z" , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . put ( "m3" 1 , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" , "m3" 5 ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" , "m3" ) ) ) . put ( "m3" 8 , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . put ( "2015-02-11T00:00:00.000Z/2015-02-12T00:00:00.000Z" , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "m3" 5 ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m3" ) ) ) . put ( "m3" 4 , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . put ( "2015-03-13T00:00:00.000Z/2015-03-19T00:00:00.000Z" , com . google . common . collect . ImmutableMap . of ( org . apache . druid . server . ClientInfoResourceTest . KEY_DIMENSIONS , com . google . common . collect . ImmutableSet . of ( "d1" ) , org . apache . druid . server . ClientInfoResourceTest . KEY_METRICS , com . google . common . collect . ImmutableSet . of ( "m1" ) ) ) . build ( ) ; org . easymock . EasyMock . verify ( serverInventoryView , timelineServerView ) ; "<AssertPlaceHolder>" ; } verify ( java . lang . Iterable , java . lang . Iterable ) { java . util . Iterator < org . apache . druid . query . Result < org . apache . druid . query . select . SelectResultValue > > expectedIter = expectedResults . iterator ( ) ; java . util . Iterator < org . apache . druid . query . Result < org . apache . druid . query . select . SelectResultValue > > actualIter =
org . junit . Assert . assertEquals ( expected , actual )
testAutoLongKeys ( ) { com . google . appengine . api . datastore . Key parent = com . dotweblabs . twirl . object . KeyStructure . createKey ( "TestKind" , com . dotweblabs . twirl . object . KeyStructure . autoLongId ( "TestKind" ) ) ; "<AssertPlaceHolder>" ; } autoLongId ( java . lang . String ) { return com . google . appengine . api . datastore . DatastoreServiceFactory . getDatastoreService ( ) . allocateIds ( kind , 1 ) . getStart ( ) . getId ( ) ; }
org . junit . Assert . assertNotNull ( parent )
testSubEpochs ( ) { System . out . println ( "getSubEpochs" ) ; jsat . classifiers . svm . extended . AMM instance = new jsat . classifiers . svm . extended . AMM ( ) ; instance . setSubEpochs ( 10 ) ; "<AssertPlaceHolder>" ; for ( int i = - 3 ; i < 1 ; i ++ ) try { instance . setSubEpochs ( i ) ; org . junit . Assert . fail ( "Invalid<sp>value<sp>should<sp>have<sp>thrown<sp>an<sp>error" ) ; } catch ( java . lang . Exception ex ) { } } getSubEpochs ( ) { return subEpochs ; }
org . junit . Assert . assertEquals ( 10 , instance . getSubEpochs ( ) )
testGetLoadSchemasCatalogs ( ) { classUnderTest . setLoadSchemasCatalogs ( true ) ; "<AssertPlaceHolder>" ; } getLoadSchemasCatalogs ( ) { return _loadSchemasCatalogs ; }
org . junit . Assert . assertEquals ( true , classUnderTest . getLoadSchemasCatalogs ( ) )
shouldSerializeGenericClass ( ) { java . lang . String expectedResult = "{\"genericWrapper\":{\"entityList\":[{\"name\":\"washington<sp>botelho\"},{\"name\":\"washington<sp>botelho\"}],\"total\":2}}" ; java . util . Collection < br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . Client > entityList = new java . util . ArrayList < br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . Client > ( ) ; entityList . add ( new br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . Client ( "washington<sp>botelho" ) ) ; entityList . add ( new br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . Client ( "washington<sp>botelho" ) ) ; br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . GenericWrapper < br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . Client > wrapper = new br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . GenericWrapper < br . com . caelum . vraptor . serialization . gson . GsonJSONSerializationTest . Client > ( entityList , entityList . size ( ) ) ; serialization . from ( wrapper ) . include ( "entityList" ) . serialize ( ) ; "<AssertPlaceHolder>" ; } result ( ) { return new java . lang . String ( stream . toByteArray ( ) ) ; }
org . junit . Assert . assertThat ( result ( ) , org . hamcrest . Matchers . is ( org . hamcrest . Matchers . equalTo ( expectedResult ) ) )
wrapTest ( ) { cn . hutool . core . io . IORuntimeException e = cn . hutool . core . exceptions . ExceptionUtil . wrap ( new java . io . IOException ( ) , cn . hutool . core . io . IORuntimeException . class ) ; "<AssertPlaceHolder>" ; } wrap ( java . lang . Throwable , java . lang . Class ) { if ( wrapThrowable . isInstance ( throwable ) ) { return ( ( T ) ( throwable ) ) ; } return cn . hutool . core . util . ReflectUtil . newInstance ( wrapThrowable , throwable ) ; }
org . junit . Assert . assertNotNull ( e )
givenHashtable_whenModifyUnderlyingCollection_thenItHasNoEffectOnIteratedCollection ( ) { java . util . Hashtable < java . lang . String , java . lang . String > table = new java . util . Hashtable < java . lang . String , java . lang . String > ( ) ; table . put ( "key1" , "value1" ) ; table . put ( "key2" , "value2" ) ; java . util . List < java . lang . String > keysSelected = com . google . common . collect . Lists . newArrayList ( ) ; java . util . Enumeration < java . lang . String > keys = table . keys ( ) ; while ( keys . hasMoreElements ( ) ) { java . lang . String key = keys . nextElement ( ) ; keysSelected . add ( key ) ; if ( key . equals ( "key1" ) ) { table . put ( "key3" , "value3" ) ; } } "<AssertPlaceHolder>" ; } size ( ) { return elements . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , keysSelected . size ( ) )
shouldHandleCharacterWhenCaseSensitive ( ) { content = "<sp>and" ; makeCaseSensitive ( ) ; tokens . consume ( "" ) ; tokens . consume ( "and" ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { return ( index ) < ( size ) ; }
org . junit . Assert . assertThat ( tokens . hasNext ( ) , org . hamcrest . core . Is . is ( false ) )
testExecuteMetGewijzigdeCategorieen ( ) { final java . util . Map < java . lang . String , java . lang . Object > parameters = new java . util . HashMap ( ) ; final nl . bzk . migratiebrp . bericht . model . lo3 . impl . Tb02Bericht bericht = tb02Factory . maakTb02Bericht ( Tb02Factory . Soort . OMZETTING_INCORRECT ) ; bericht . setHeader ( Lo3HeaderVeld . AKTENUMMER , "3QH5432" ) ; parameters . put ( "input" , berichtenDao . bewaarBericht ( bericht ) ) ; final java . util . Map < java . lang . String , java . lang . Object > result = subject . execute ( parameters ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return elementen . isEmpty ( ) ; }
org . junit . Assert . assertFalse ( result . isEmpty ( ) )
test8 ( ) { com . creactiviti . piper . core . MapObject mo = new com . creactiviti . piper . core . MapObject ( java . util . Collections . singletonMap ( "hello" , "world" ) ) { } ; "<AssertPlaceHolder>" ; } get ( javax . servlet . http . HttpServletRequest ) { java . lang . String path = ( ( java . lang . String ) ( aRequest . getAttribute ( HandlerMapping . PATH_WITHIN_HANDLER_MAPPING_ATTRIBUTE ) ) ) ; java . lang . String pipelineId = path . replaceFirst ( "/pipelines/" , "" ) ; return pipelineRepository . findOne ( pipelineId ) ; }
org . junit . Assert . assertEquals ( "world" , mo . get ( "hello" ) )
testEmptyItemEquals ( ) { com . vaadin . v7 . data . util . PropertysetItem item1 = createPropertySetItem ( ) ; com . vaadin . v7 . data . util . PropertysetItem item2 = createPropertySetItem ( ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) { return true ; } else if ( obj instanceof com . vaadin . client . metadata . Property ) { com . vaadin . client . metadata . Property other = ( ( com . vaadin . client . metadata . Property ) ( obj ) ) ; return getSignature ( ) . equals ( other . getSignature ( ) ) ; } else { return false ; } }
org . junit . Assert . assertTrue ( item1 . equals ( item2 ) )
testRecommendPatterns_secondPattern_twoDates ( ) { java . util . List < java . lang . String [ ] > patternCombinations = org . esa . snap . opendap . utils . PatternProvider . recommendPatterns ( "sth__20100101_192345_20110101_192345.nc" ) ; "<AssertPlaceHolder>" ; org . esa . snap . opendap . utils . PatternProviderTest . containAssertEquals ( patternCombinations , new java . lang . String [ ] { "yyyyMMdd_hhmmss" , "*${startDate}*" } ) ; org . esa . snap . opendap . utils . PatternProviderTest . containAssertEquals ( patternCombinations , new java . lang . String [ ] { "yyyyMMdd_hhmmss" , "*${startDate}*${endDate}*" } ) ; org . esa . snap . opendap . utils . PatternProviderTest . containAssertEquals ( patternCombinations , new java . lang . String [ ] { "yyyyMMdd" , "*${startDate}*" } ) ; org . esa . snap . opendap . utils . PatternProviderTest . containAssertEquals ( patternCombinations , new java . lang . String [ ] { "yyyyMMdd" , "*${startDate}*${endDate}*" } ) ; org . esa . snap . opendap . utils . PatternProviderTest . containAssertEquals ( patternCombinations , new java . lang . String [ ] { "yyyyMM" , "*${startDate}*" } ) ; org . esa . snap . opendap . utils . PatternProviderTest . containAssertEquals ( patternCombinations , new java . lang . String [ ] { "yyyyMM" , "*${startDate}*${endDate}*" } ) ; } recommendPatterns ( java . lang . String ) { fileName = org . esa . snap . opendap . utils . PatternProvider . replaceColon ( fileName ) ; java . util . List < java . lang . String [ ] > patternCombinations = new java . util . ArrayList < java . lang . String [ ] > ( ) ; for ( java . lang . String dateString : org . esa . snap . opendap . utils . PatternProvider . DATE_PATTERNS ) { dateString = org . esa . snap . opendap . utils . PatternProvider . replaceColon ( dateString ) ; final java . util . regex . Pattern datePattern = org . esa . snap . opendap . utils . PatternProvider . convertDateStringToPattern ( dateString ) ; java . util . regex . Matcher matcher = datePattern . matcher ( fileName ) ; if ( matcher . matches ( ) ) { org . esa . snap . util . TimeStampExtractor timeStampExtractor = new org . esa . snap . util . TimeStampExtractor ( dateString , org . esa . snap . opendap . utils . PatternProvider . FILENAME_PATTERNS [ 0 ] ) ; try { timeStampExtractor . extractTimeStamps ( fileName ) ; patternCombinations . add ( new java . lang . String [ ] { org . esa . snap . opendap . utils . PatternProvider . insertColon ( dateString ) , org . esa . snap . opendap . utils . PatternProvider . FILENAME_PATTERNS [ 0 ] } ) ; try { timeStampExtractor = new org . esa . snap . util . TimeStampExtractor ( org . esa . snap . opendap . utils . PatternProvider . insertColon ( dateString ) , org . esa . snap . opendap . utils . PatternProvider . FILENAME_PATTERNS [ 1 ] ) ; timeStampExtractor . extractTimeStamps ( fileName ) ; patternCombinations . add ( new java . lang . String [ ] { dateString , org . esa . snap . opendap . utils . PatternProvider . FILENAME_PATTERNS [ 1 ] } ) ; } catch ( com . bc . ceres . binding . ValidationException e ) { } } catch ( com . bc . ceres . binding . ValidationException e ) { } } } return patternCombinations ; }
org . junit . Assert . assertTrue ( ( ( patternCombinations . size ( ) ) >= 6 ) )
getDetails ( ) { final java . lang . String instanceName = getRyaInstanceName ( ) ; final org . apache . rya . api . client . Install . InstallConfiguration installConfig = org . apache . rya . api . client . Install . InstallConfiguration . builder ( ) . setEnableTableHashPrefix ( true ) . setEnableEntityCentricIndex ( true ) . setEnableFreeTextIndex ( true ) . setEnableTemporalIndex ( true ) . setEnablePcjIndex ( true ) . setEnableGeoIndex ( true ) . build ( ) ; final org . apache . rya . api . client . accumulo . AccumuloConnectionDetails connectionDetails = new org . apache . rya . api . client . accumulo . AccumuloConnectionDetails ( getUsername ( ) , getPassword ( ) . toCharArray ( ) , getInstanceName ( ) , getZookeepers ( ) ) ; final org . apache . rya . api . client . Install install = new org . apache . rya . api . client . accumulo . AccumuloInstall ( connectionDetails , getConnector ( ) ) ; install . install ( instanceName , installConfig ) ; final org . apache . rya . api . client . GetInstanceDetails getInstanceDetails = new org . apache . rya . api . client . accumulo . AccumuloGetInstanceDetails ( connectionDetails , getConnector ( ) ) ; final com . google . common . base . Optional < org . apache . rya . api . instance . RyaDetails > details = getInstanceDetails . getDetails ( instanceName ) ; final org . apache . rya . api . instance . RyaDetails expectedDetails = . setGeoIndexDetails ( new GeoIndexDetails ( true ) ) org . apache . rya . api . instance . RyaDetails . builder ( ) . setRyaInstanceName ( instanceName ) . setRyaVersion ( details . get ( ) . getRyaVersion ( ) ) . setTemporalIndexDetails ( new org . apache . rya . api . instance . RyaDetails . TemporalIndexDetails ( true ) ) . setFreeTextDetails ( new org . apache . rya . api . instance . RyaDetails . FreeTextIndexDetails ( true ) ) . setEntityCentricIndexDetails ( new org . apache . rya . api . instance . RyaDetails . EntityCentricIndexDetails ( true ) ) . setPCJIndexDetails ( org . apache . rya . api . instance . RyaDetails . PCJIndexDetails . builder ( ) . setEnabled ( true ) ) . setProspectorDetails ( new org . apache . rya . api . instance . RyaDetails . ProspectorDetails ( com . google . common . base . Optional . < java . util . Date > absent ( ) ) ) . setJoinSelectivityDetails ( new org . apache . rya . api . instance . RyaDetails . JoinSelectivityDetails ( com . google . common . base . Optional . < java . util . Date > absent ( ) ) ) . build ( ) ; "<AssertPlaceHolder>" ; } get ( ) { if ( ( metadata ) == null ) { metadata = getMetadataFromTable ( ) ; } return metadata ; }
org . junit . Assert . assertEquals ( expectedDetails , details . get ( ) )
testFilteredNodeIterator ( ) { java . util . HashSet < org . graphstream . graph . Node > expected = new java . util . HashSet < org . graphstream . graph . Node > ( ) ; java . util . HashSet < org . graphstream . graph . Node > reached = new java . util . HashSet < org . graphstream . graph . Node > ( ) ; org . graphstream . util . FilteredNodeIterator < org . graphstream . graph . Node > ite = new org . graphstream . util . FilteredNodeIterator < org . graphstream . graph . Node > ( baseGraph . getNodeIterator ( ) , org . graphstream . util . Filters . < org . graphstream . graph . Node > byAttributeFilter ( "type" , "A" ) ) ; expected . add ( baseGraph . getNode ( "A0" ) ) ; expected . add ( baseGraph . getNode ( "A1" ) ) ; expected . add ( baseGraph . getNode ( "A2" ) ) ; while ( ite . hasNext ( ) ) reached . add ( ite . next ( ) ) ; "<AssertPlaceHolder>" ; } next ( ) { if ( ( iNext ) >= ( iEnd ) ) throw new java . util . NoSuchElementException ( ) ; iPrev = ( iNext ) ++ ; return ( ( T ) ( edges [ iPrev ] ) ) ; }
org . junit . Assert . assertEquals ( expected , reached )
testElasticIndex ( ) { com . msd . gin . halyard . sail . HBaseSailConfig cfg = new com . msd . gin . halyard . sail . HBaseSailConfig ( ) ; cfg . setElasticIndexURL ( "http://localhost:12345/index" ) ; "<AssertPlaceHolder>" ; } getElasticIndexURL ( ) { return elasticIndexURL ; }
org . junit . Assert . assertEquals ( "http://localhost:12345/index" , cfg . getElasticIndexURL ( ) )
shouldConsiderChildPathToNotBeAncestorOfParent ( ) { "<AssertPlaceHolder>" ; } isAncestorOf ( org . modeshape . jcr . value . Path ) { org . modeshape . common . util . CheckArg . isNotNull ( descendant , "descendant" ) ; return descendant . isDescendantOf ( this ) ; }
org . junit . Assert . assertThat ( path . isAncestorOf ( parent ) , org . hamcrest . core . Is . is ( false ) )
httpRequestWasMade_returnsTrueIfRequestMatchingGivenRuleWasMade ( ) { makeRequest ( "http://example.com" ) ; "<AssertPlaceHolder>" ; } httpRequestWasMade ( java . lang . String ) { return org . robolectric . shadows . httpclient . FakeHttp . getFakeHttpLayer ( ) . hasRequestMatchingRule ( new org . robolectric . shadows . httpclient . FakeHttpLayer . UriRequestMatcher ( uri ) ) ; }
org . junit . Assert . assertTrue ( org . robolectric . shadows . httpclient . FakeHttp . httpRequestWasMade ( "http://example.com" ) )
testGetEmrClusterDefinitionEntity ( ) { org . finra . herd . model . jpa . NamespaceEntity namespaceEntity = new org . finra . herd . model . jpa . NamespaceEntity ( ) ; org . finra . herd . model . jpa . EmrClusterDefinitionEntity emrClusterDefinitionEntity = new org . finra . herd . model . jpa . EmrClusterDefinitionEntity ( ) ; org . finra . herd . model . api . xml . EmrClusterDefinitionKey emrClusterDefinitionKey = new org . finra . herd . model . api . xml . EmrClusterDefinitionKey ( NAMESPACE , EMR_CLUSTER_DEFINITION_NAME ) ; when ( namespaceDao . getNamespaceByCd ( org . finra . herd . service . helper . NAMESPACE ) ) . thenReturn ( namespaceEntity ) ; when ( emrClusterDefinitionDao . getEmrClusterDefinitionByNamespaceAndName ( namespaceEntity , org . finra . herd . service . helper . EMR_CLUSTER_DEFINITION_NAME ) ) . thenReturn ( emrClusterDefinitionEntity ) ; org . finra . herd . model . jpa . EmrClusterDefinitionEntity result = emrClusterDefinitionDaoHelper . getEmrClusterDefinitionEntity ( emrClusterDefinitionKey ) ; verify ( namespaceDao ) . getNamespaceByCd ( org . finra . herd . service . helper . NAMESPACE ) ; verify ( emrClusterDefinitionDao ) . getEmrClusterDefinitionByNamespaceAndName ( namespaceEntity , org . finra . herd . service . helper . EMR_CLUSTER_DEFINITION_NAME ) ; verifyNoMoreInteractionsHelper ( ) ; "<AssertPlaceHolder>" ; } verifyNoMoreInteractionsHelper ( ) { verifyNoMoreInteractions ( awsHelper , javaPropertiesHelper , retryPolicyFactory , s3Operations ) ; }
org . junit . Assert . assertEquals ( emrClusterDefinitionEntity , result )
testTwoNull ( ) { org . openscience . cdk . tools . diff . tree . IDifference result = org . openscience . cdk . tools . diff . tree . BooleanArrayDifference . construct ( "Foo" , null , null ) ; "<AssertPlaceHolder>" ; } construct ( java . lang . String , boolean [ ] , boolean [ ] ) { if ( ( first == null ) && ( second == null ) ) return null ; org . openscience . cdk . tools . diff . tree . BooleanArrayDifference totalDiff = new org . openscience . cdk . tools . diff . tree . BooleanArrayDifference ( name ) ; int firstLength = ( first == null ) ? 0 : first . length ; int secondLength = ( second == null ) ? 0 : second . length ; if ( firstLength == secondLength ) { for ( int i = 0 ; i < firstLength ; i ++ ) { totalDiff . addChild ( org . openscience . cdk . tools . diff . tree . BooleanDifference . construct ( ( "" + i ) , first [ i ] , second [ i ] ) ) ; } } else if ( firstLength < secondLength ) { for ( int i = 0 ; i < firstLength ; i ++ ) { totalDiff . addChild ( org . openscience . cdk . tools . diff . tree . BooleanDifference . construct ( ( "" + i ) , first [ i ] , second [ i ] ) ) ; } for ( int i = firstLength ; i < secondLength ; i ++ ) { totalDiff . addChild ( org . openscience . cdk . tools . diff . tree . BooleanDifference . construct ( ( "" + i ) , null , second [ i ] ) ) ; } } else { for ( int i = 0 ; i < secondLength ; i ++ ) { totalDiff . addChild ( org . openscience . cdk . tools . diff . tree . BooleanDifference . construct ( ( "" + i ) , first [ i ] , second [ i ] ) ) ; } for ( int i = secondLength ; i < firstLength ; i ++ ) { totalDiff . addChild ( org . openscience . cdk . tools . diff . tree . BooleanDifference . construct ( ( "" + i ) , first [ i ] , null ) ) ; } } if ( ( totalDiff . childCount ( ) ) == 0 ) { return null ; } return totalDiff ; }
org . junit . Assert . assertNull ( result )
testToNullWidgetValue ( ) { java . lang . Object value = converter . toWidgetValue ( null ) ; "<AssertPlaceHolder>" ; } toWidgetValue ( java . lang . Integer ) { return modelValue != null ? modelValue . doubleValue ( ) : null ; }
org . junit . Assert . assertNull ( value )
multipleUrlResultParsing ( ) { doc = com . rosaloves . bitlyj . Utils . classpathXmlIS ( "clicks_2_urls.xml" ) ; java . util . Set < com . rosaloves . bitlyj . UrlClicks > clicks = com . rosaloves . bitlyj . Bitly . clicks ( "http://tcrn.ch/a4MSUH" , "http://bit.ly/1YKMfY" ) . apply ( bitly , doc ) ; "<AssertPlaceHolder>" ; } size ( ) { return parameters . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , clicks . size ( ) )
shouldNotEquals ( ) { byte [ ] bytes = "string" . getBytes ( ) ; co . nstant . in . cbor . model . ByteString byteString = new co . nstant . in . cbor . model . ByteString ( bytes ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( object instanceof co . nstant . in . cbor . model . UnicodeString ) && ( super . equals ( object ) ) ) { co . nstant . in . cbor . model . UnicodeString other = ( ( co . nstant . in . cbor . model . UnicodeString ) ( object ) ) ; if ( ( string ) == null ) { return ( other . string ) == null ; } else { return string . equals ( other . string ) ; } } return false ; }
org . junit . Assert . assertFalse ( byteString . equals ( new java . lang . Object ( ) ) )
testLenientParsing ( ) { final org . opendaylight . yangtools . yang . model . api . SchemaContext schemaContext = org . opendaylight . yangtools . yang . test . util . YangParserTestUtils . parseYangResource ( "/strict-parsing-mode-test/foo.yang" ) ; final org . opendaylight . yangtools . yang . model . api . Module fooModule = schemaContext . getModules ( ) . iterator ( ) . next ( ) ; final org . opendaylight . yangtools . yang . model . api . ContainerSchemaNode topLevelContainer = ( ( org . opendaylight . yangtools . yang . model . api . ContainerSchemaNode ) ( fooModule . findDataChildByName ( org . opendaylight . yangtools . yang . common . QName . create ( fooModule . getQNameModule ( ) , "top-level-container" ) ) . get ( ) ) ) ; final java . io . InputStream resourceAsStream = org . opendaylight . yangtools . yang . data . codec . xml . StrictParsingModeTest . class . getResourceAsStream ( "/strict-parsing-mode-test/foo.xml" ) ; final javax . xml . stream . XMLStreamReader reader = org . opendaylight . yangtools . util . xml . UntrustedXML . createXMLStreamReader ( resourceAsStream ) ; final org . opendaylight . yangtools . yang . data . impl . schema . NormalizedNodeResult result = new org . opendaylight . yangtools . yang . data . impl . schema . NormalizedNodeResult ( ) ; final org . opendaylight . yangtools . yang . data . api . schema . stream . NormalizedNodeStreamWriter streamWriter = org . opendaylight . yangtools . yang . data . impl . schema . ImmutableNormalizedNodeStreamWriter . from ( result ) ; final org . opendaylight . yangtools . yang . data . codec . xml . XmlParserStream xmlParser = org . opendaylight . yangtools . yang . data . codec . xml . XmlParserStream . create ( streamWriter , schemaContext , topLevelContainer , false ) ; xmlParser . parse ( reader ) ; final org . opendaylight . yangtools . yang . data . api . schema . NormalizedNode < ? , ? > transformedInput = result . getResult ( ) ; "<AssertPlaceHolder>" ; } getResult ( ) { checkState ( ( ( result ) != null ) , "Metadata<sp>tree<sp>has<sp>not<sp>been<sp>completely<sp>built" ) ; return result . getAnnotations ( ) . isEmpty ( ) ? java . util . Optional . empty ( ) : java . util . Optional . of ( result ) ; }
org . junit . Assert . assertNotNull ( transformedInput )
testMapStringReturn ( ) { io . vertx . rx . java . test . gen . Map < java . lang . String , java . lang . String > map = obj . methodWithMapStringReturn ( ( s ) -> { } ) ; "<AssertPlaceHolder>" ; } get ( examples . HttpClient ) { examples . Single < examples . HttpClientResponse > get = client . rxGetNow ( "http://the-server" ) ; get . subscribe ( ( resp ) -> { } , ( err ) -> { } ) ; }
org . junit . Assert . assertEquals ( "bar" , map . get ( "foo" ) )
xStream11Compatibility ( ) { hudson . util . XStream2Test . Bar b = ( ( hudson . util . XStream2Test . Bar ) ( new hudson . util . XStream2 ( ) . fromXML ( "<hudson.util.XStream2Test-Bar><s>foo</s></hudson.util.XStream2Test-Bar>" ) ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( "foo" , b . s )
testFromEbXMLObjRef ( ) { org . openehealth . ipf . commons . ihe . xds . core . ebxml . EbXMLQueryResponse ebXML = transformer . toEbXML ( responseObjRef ) ; org . openehealth . ipf . commons . ihe . xds . core . responses . QueryResponse result = transformer . fromEbXML ( ebXML ) ; "<AssertPlaceHolder>" ; } fromEbXML ( org . openehealth . ipf . commons . ihe . xds . core . ebxml . EbXMLRetrieveDocumentSetRequest ) { if ( ebXML == null ) { return null ; } org . openehealth . ipf . commons . ihe . xds . core . requests . RetrieveDocumentSet request = new org . openehealth . ipf . commons . ihe . xds . core . requests . RetrieveDocumentSet ( ) ; request . getDocuments ( ) . addAll ( ebXML . getDocuments ( ) ) ; return request ; }
org . junit . Assert . assertEquals ( responseObjRef , result )
appendsAccessTokenToBody ( ) { java . lang . String expectedBodyContent = "access_token=" + ( profileWithOnlyAccessToken . getAccessToken ( ) ) ; profileWithOnlyAccessToken . setAccessTokenPosition ( OAuth2Profile . AccessTokenPosition . BODY ) ; oltuClientFacade . applyAccessToken ( profileWithOnlyAccessToken , httpRequest , "" ) ; java . io . StringWriter writer = new java . io . StringWriter ( ) ; org . apache . commons . io . IOUtils . copy ( httpRequest . getEntity ( ) . getContent ( ) , writer , "UTF-8" ) ; java . lang . String actualContent = writer . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return getURLDecodedName ( ) ; }
org . junit . Assert . assertThat ( actualContent , org . hamcrest . CoreMatchers . is ( expectedBodyContent ) )
testMapKeyValuePartial ( ) { one . util . streamex . Map < java . lang . Integer , java . lang . Integer > original = new one . util . streamex . HashMap ( ) ; original . put ( 1 , 1 ) ; original . put ( 2 , 5 ) ; original . put ( 3 , 3 ) ; original . put ( 4 , 4 ) ; one . util . streamex . List < java . lang . Integer > expected = asList ( 1 , 9 , 16 ) ; one . util . streamex . List < java . lang . Integer > actual = one . util . streamex . EntryStream . of ( original ) . mapKeyValuePartial ( ( key , value ) -> { if ( key . equals ( value ) ) { return one . util . streamex . Optional . of ( ( key * value ) ) ; } return one . util . streamex . Optional . empty ( ) ; } ) . toList ( ) ; "<AssertPlaceHolder>" ; } empty ( ) { return one . util . streamex . DoubleStreamEx . of ( one . util . streamex . Spliterators . emptyDoubleSpliterator ( ) ) ; }
org . junit . Assert . assertEquals ( expected , actual )
testLogicalLogPrepared ( ) { javax . transaction . xa . Xid xid = new org . neo4j . kernel . impl . transaction . XidImpl ( new byte [ 2 ] , new byte [ 2 ] ) ; javax . transaction . xa . XAResource xaRes = xaCon . getXaResource ( ) ; xaRes . start ( xid , XAResource . TMNOFLAGS ) ; long node1 = ds . nextId ( org . neo4j . graphdb . Node . class ) ; xaCon . getNodeConsumer ( ) . createNode ( node1 ) ; long node2 = ds . nextId ( org . neo4j . graphdb . Node . class ) ; xaCon . getNodeConsumer ( ) . createNode ( node2 ) ; long n1prop1 = ds . nextId ( org . neo4j . kernel . impl . nioneo . store . PropertyStore . class ) ; xaCon . getNodeConsumer ( ) . addProperty ( node1 , n1prop1 , index ( "prop1" ) , "string1" ) ; int relType1 = ( ( int ) ( ds . nextId ( org . neo4j . graphdb . RelationshipType . class ) ) ) ; xaCon . getRelationshipTypeConsumer ( ) . addRelationshipType ( relType1 , "relationshiptype1" ) ; long rel1 = ds . nextId ( org . neo4j . graphdb . Relationship . class ) ; xaCon . getRelationshipConsumer ( ) . createRelationship ( rel1 , node1 , node2 , relType1 ) ; long r1prop1 = ds . nextId ( org . neo4j . kernel . impl . nioneo . store . PropertyStore . class ) ; xaCon . getRelationshipConsumer ( ) . addProperty ( rel1 , r1prop1 , index ( "prop1" ) , "string1" ) ; xaCon . getNodeConsumer ( ) . changeProperty ( node1 , n1prop1 , "string2" ) ; xaCon . getRelationshipConsumer ( ) . changeProperty ( rel1 , r1prop1 , "string2" ) ; xaRes . end ( xid , XAResource . TMSUCCESS ) ; xaRes . prepare ( xid ) ; ds . rotateLogicalLog ( ) ; org . neo4j . kernel . impl . nioneo . store . TestXa . copyLogicalLog ( path ( ) ) ; xaCon . clearAllTransactions ( ) ; ds . close ( ) ; deleteLogicalLogIfExist ( ) ; org . neo4j . kernel . impl . nioneo . store . TestXa . renameCopiedLogicalLog ( path ( ) ) ; ds = newNeoStore ( ) ; xaCon = ( ( org . neo4j . kernel . impl . nioneo . xa . NeoStoreXaConnection ) ( ds . getXaConnection ( ) ) ) ; xaRes = xaCon . getXaResource ( ) ; "<AssertPlaceHolder>" ; xaRes . commit ( xid , true ) ; xaCon . clearAllTransactions ( ) ; } recover ( java . util . Iterator ) { msgLog . logMessage ( ( "<sp>transactions<sp>already<sp>rolled<sp>back." 4 + ( txLog . getName ( ) ) ) , true ) ; try { java . util . List < org . neo4j . kernel . impl . transaction . TxManager . NonCompletedTransaction > commitList = new java . util . ArrayList < org . neo4j . kernel . impl . transaction . TxManager . NonCompletedTransaction > ( ) ; java . util . List < javax . transaction . xa . Xid > rollbackList = new java . util . LinkedList < javax . transaction . xa . Xid > ( ) ; java . util . Map < org . neo4j . kernel . impl . transaction . TxManager . Resource , javax . transaction . xa . XAResource > resourceMap = new java . util . HashMap < org . neo4j . kernel . impl . transaction . TxManager . Resource , javax . transaction . xa . XAResource > ( ) ; buildRecoveryInfo ( commitList , rollbackList , resourceMap , danglingRecordList ) ; java . util . Iterator < org . neo4j . kernel . impl . transaction . TxManager . Resource > resourceItr = resourceMap . keySet ( ) . iterator ( ) ; java . util . List < javax . transaction . xa . Xid > recoveredXidsList = new java . util . LinkedList < javax . transaction . xa . Xid > ( ) ; while ( resourceItr . hasNext ( ) ) { javax . transaction . xa . XAResource xaRes = resourceMap . get ( resourceItr . next ( ) ) ; javax . transaction . xa . Xid [ ] xids = xaRes . recover ( XAResource . TMNOFLAGS ) ; for ( int i = 0 ; i < ( xids . length ) ; i ++ ) { if ( org . neo4j . kernel . impl . transaction . XidImpl . isThisTm ( xids [ i ] . getGlobalTransactionId ( ) ) ) { if ( rollbackList . contains ( xids [ i ] ) ) { org . neo4j . kernel . impl . transaction . TxManager . log . fine ( ( ( "Unknown<sp>xid:<sp>" 8 + ( xids [ i ] ) ) + "<sp>rolling<sp>back<sp>...<sp>" ) ) ; msgLog . logMessage ( ( ( "TM:<sp>Found<sp>pre<sp>commit<sp>" + ( xids [ i ] ) ) + "<sp>rolling<sp>back<sp>...<sp>" ) , true ) ; rollbackList . remove ( xids [ i ] ) ; xaRes . rollback ( xids [ i ] ) ; } else { recoveredXidsList . add ( xids [ i ] ) ; } } else { org . neo4j . kernel . impl . transaction . TxManager . log . warning ( ( "Unknown<sp>xid:<sp>" + ( xids [ i ] ) ) ) ; } } } java . util . Collections . sort ( commitList , new java . util . Comparator < org . neo4j . kernel . impl . transaction . TxManager . NonCompletedTransaction > ( ) { public int compare ( org . neo4j . kernel . impl . transaction . TxManager . NonCompletedTransaction r1 , org . neo4j . kernel . impl . transaction . TxManager
org . junit . Assert . assertEquals ( 1 , xaRes . recover ( XAResource . TMNOFLAGS ) . length )
getFourth_A$ ( ) { java . lang . String _1 = "foo" ; java . lang . Integer _2 = 123 ; java . lang . Long _3 = 456L ; java . lang . Boolean _4 = true ; java . lang . Double _5 = 7.8 ; com . m3 . scalaflavor4j . Tuple5 < java . lang . String , java . lang . Integer , java . lang . Long , java . lang . Boolean , java . lang . Double > tuple = com . m3 . scalaflavor4j . Tuple . apply ( _1 , _2 , _3 , _4 , _5 ) ; java . lang . Boolean actual = tuple . getFourth ( ) ; java . lang . Boolean expected = true ; "<AssertPlaceHolder>" ; } getFourth ( ) { return _4 ( ) ; }
org . junit . Assert . assertThat ( actual , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( expected ) ) )
testFetchByPrimaryKeysWithMultiplePrimaryKeysWhereNoPrimaryKeysExist ( ) { long pk1 = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; long pk2 = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; primaryKeys . add ( pk1 ) ; primaryKeys . add ( pk2 ) ; java . util . Map < java . io . Serializable , com . liferay . asset . kernel . model . AssetTag > assetTags = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( assetTags . isEmpty ( ) )
testThrowOnNullSubjectWithQueue ( ) { try ( io . nats . client . NatsTestServer ts = new io . nats . client . NatsTestServer ( false ) ; io . nats . client . Connection nc = io . nats . client . Nats . connect ( ts . getURI ( ) ) ) { io . nats . client . Dispatcher d = nc . createDispatcher ( ( msg ) -> { } ) ; d . subscribe ( null , "quque" ) ; "<AssertPlaceHolder>" ; } } subscribe ( java . lang . String , java . lang . String ) { if ( ( subject == null ) || ( ( subject . length ( ) ) == 0 ) ) { throw new java . lang . IllegalArgumentException ( "Subject<sp>is<sp>required<sp>in<sp>subscribe" ) ; } if ( ( queueName == null ) || ( ( queueName . length ( ) ) == 0 ) ) { throw new java . lang . IllegalArgumentException ( "QueueName<sp>is<sp>required<sp>in<sp>subscribe" ) ; } return createSubscription ( subject , queueName , null ) ; }
org . junit . Assert . assertFalse ( true )
testJaccardCompare2 ( ) { final double EXPECT = 0.9357849740192014 ; java . lang . String inputStr = "account" ; java . lang . String outputStr = "account<sp>status<sp>c" ; double result = org . talend . utils . string . Jaccard . JaccardCompare ( inputStr , outputStr ) ; "<AssertPlaceHolder>" ; } JaccardCompare ( java . lang . String , java . lang . String ) { java . lang . String [ ] left = res . split ( "\\s+" ) ; java . lang . String [ ] right = res1 . split ( "\\s+" ) ; int leftLength = left . length ; int rightLength = right . length ; java . util . Set < java . lang . String > unionSet = new java . util . HashSet < java . lang . String > ( ) ; boolean unionFilled = false ; double intersection = 0 ; if ( ( leftLength == 0 ) || ( rightLength == 0 ) ) { return 0.0 ; } for ( int leftIndex = 0 ; leftIndex < leftLength ; leftIndex ++ ) { unionSet . add ( left [ leftIndex ] ) ; for ( int rightIndex = 0 ; rightIndex < rightLength ; rightIndex ++ ) { if ( ! unionFilled ) { unionSet . add ( right [ rightIndex ] ) ; } if ( left [ leftIndex ] . equals ( right [ rightIndex ] ) ) { int wordLength = left [ leftIndex ] . length ( ) ; if ( wordLength > 1 ) { double weight = ( java . lang . Math . log ( wordLength ) ) / ( java . lang . Math . log ( 2 ) ) ; intersection = intersection + weight ; } else { intersection ++ ; } } } unionFilled = true ; } return intersection / ( java . lang . Double . valueOf ( unionSet . size ( ) ) ) ; }
org . junit . Assert . assertTrue ( ( result == EXPECT ) )
testGetReferencedGroups ( ) { java . lang . String [ ] expected = new java . lang . String [ ] { "item" } ; doReturn ( expected ) . when ( expressionPropertyCore ) . getReferencedGroups ( metaData , element , null ) ; java . lang . String [ ] result = metaData . getReferencedGroups ( element , null ) ; "<AssertPlaceHolder>" ; } getReferencedGroups ( org . pentaho . reporting . engine . classic . core . function . Expression , java . lang . Object ) { return expressionPropertyCore . getReferencedGroups ( this , element , attributeValue ) ; }
org . junit . Assert . assertThat ( result , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( expected ) ) )
testSshKey ( ) { org . gitlab4j . api . models . SshKey sshKey = org . gitlab4j . api . JsonUtils . unmarshalResource ( org . gitlab4j . api . models . SshKey . class , "sshkey.json" ) ; "<AssertPlaceHolder>" ; } compareJson ( T , java . lang . String ) { java . io . InputStreamReader reader = new java . io . InputStreamReader ( org . gitlab4j . api . TestGitLabApiBeans . class . getResourceAsStream ( filename ) ) ; return org . gitlab4j . api . JsonUtils . compareJson ( apiObject , reader ) ; }
org . junit . Assert . assertTrue ( org . gitlab4j . api . JsonUtils . compareJson ( sshKey , "sshkey.json" ) )
runInTransactionShouldCallFunctionAndCommit ( ) { final com . spotify . styx . storage . DatastoreStorage storage = new com . spotify . styx . storage . DatastoreStorage ( datastore , java . time . Duration . ZERO , storageTransactionFactory , executor ) ; final com . spotify . styx . storage . CheckedDatastoreTransaction transaction = datastore . newTransaction ( ) ; final com . spotify . styx . storage . DatastoreStorageTransaction storageTransaction = spy ( new com . spotify . styx . storage . DatastoreStorageTransaction ( transaction ) ) ; when ( storageTransactionFactory . apply ( any ( ) ) ) . thenReturn ( storageTransaction ) ; when ( transactionFunction . apply ( any ( ) ) ) . thenReturn ( "foo" ) ; java . lang . String result = storage . runInTransaction ( transactionFunction ) ; "<AssertPlaceHolder>" ; verify ( transactionFunction ) . apply ( storageTransaction ) ; verify ( storageTransaction ) . commit ( ) ; verify ( storageTransaction , never ( ) ) . rollback ( ) ; } is ( com . spotify . styx . api . Api$Version ) { return new org . hamcrest . TypeSafeMatcher < com . spotify . styx . api . Api . Version > ( ) { @ com . spotify . styx . api . Override protected boolean matchesSafely ( com . spotify . styx . api . Api . Version item ) { return ( item . ordinal ( ) ) == ( version . ordinal ( ) ) ; } @ com . spotify . styx . api . Override public void describeTo ( org . hamcrest . Description description ) { description . appendText ( "Version<sp>can<sp>only<sp>be" ) ; description . appendValue ( version ) ; } } ; }
org . junit . Assert . assertThat ( result , org . hamcrest . Matchers . is ( "foo" ) )
testState ( ) { executor . assign ( "two" , "2" , message ) ; executor . assign ( "four" , "4" , message ) ; executor . assign ( "sum" , "two<sp>+<sp>four" , message ) ; java . lang . Object var = executor . getState ( ) . get ( "sum" ) ; "<AssertPlaceHolder>" ; } get ( org . apache . hadoop . hbase . client . Get ) { if ( ! ( data . containsKey ( get . getRow ( ) ) ) ) return new org . apache . hadoop . hbase . client . Result ( ) ; byte [ ] row = get . getRow ( ) ; java . util . List < org . apache . hadoop . hbase . KeyValue > kvs = new java . util . ArrayList < org . apache . hadoop . hbase . KeyValue > ( ) ; if ( ! ( get . hasFamilies ( ) ) ) { kvs = org . apache . metron . hbase . mock . MockHTable . toKeyValue ( row , data . get ( row ) , get . getMaxVersions ( ) ) ; } else { for ( byte [ ] family : get . getFamilyMap ( ) . keySet ( ) ) { if ( ( data . get ( row ) . get ( family ) ) == null ) continue ; java . util . NavigableSet < byte [ ] > qualifiers = get . getFamilyMap ( ) . get ( family ) ; if ( ( qualifiers == null ) || ( qualifiers . isEmpty ( ) ) ) qualifiers = data . get ( row ) . get ( family ) . navigableKeySet ( ) ; for ( byte [ ] qualifier : qualifiers ) { if ( qualifier == null ) qualifier = "" . getBytes ( ) ; if ( ( ( ! ( data . get ( row ) . containsKey ( family ) ) ) || ( ! ( data . get ( row ) . get ( family ) . containsKey ( qualifier ) ) ) ) || ( data . get ( row ) . get ( family ) . get ( qualifier ) . isEmpty ( ) ) ) continue ; Map . Entry < java . lang . Long , byte [ ] > timestampAndValue = data . get ( row ) . get ( family ) . get ( qualifier ) . lastEntry ( ) ; kvs . add ( new org . apache . hadoop . hbase . KeyValue ( row , family , qualifier , timestampAndValue . getKey ( ) , timestampAndValue . getValue ( ) ) ) ; } } } org . apache . hadoop . hbase . filter . Filter filter = get . getFilter ( ) ; if ( filter != null ) { filter . reset ( ) ; java . util . List < org . apache . hadoop . hbase . KeyValue > nkvs = new java . util . ArrayList < org . apache . hadoop . hbase . KeyValue > ( kvs . size ( ) ) ; for ( org . apache . hadoop . hbase . KeyValue kv : kvs ) { if ( filter . filterAllRemaining ( ) ) { break ; } if ( filter . filterRowKey ( kv . getBuffer ( ) , kv . getRowOffset ( ) , kv . getRowLength ( ) ) ) { continue ; } if ( ( filter . filterKeyValue ( kv ) ) == ( Filter . ReturnCode . INCLUDE ) ) { nkvs . add ( kv ) ; } } if ( filter . hasFilterRow ( ) ) { filter . filterRow ( ) ; } kvs = nkvs ; } return new org . apache . hadoop . hbase . client . Result ( kvs ) ; }
org . junit . Assert . assertEquals ( 6 , var )
visitFreeFormLine ( ) { visitor . visit ( new org . drools . workbench . models . datamodel . rule . FreeFormLine ( ) ) ; "<AssertPlaceHolder>" ; } getRHSClasses ( ) { return classes ; }
org . junit . Assert . assertTrue ( visitor . getRHSClasses ( ) . isEmpty ( ) )
testGetContentType ( ) { ninja . Result result = new ninja . Result ( 200 ) ; result . contentType ( "text/my-funky-content-type" ) ; "<AssertPlaceHolder>" ; } getContentType ( ) { return contentType ; }
org . junit . Assert . assertEquals ( "text/my-funky-content-type" , result . getContentType ( ) )
t04_oneToOne ( ) { com . jajja . jorm . Transaction moria = psql . Moria . open ( ) ; try { moria . Goblin goblin = moria . select ( moria . Goblin . class , "SELECT<sp>*<sp>FROM<sp>#1#<sp>WHERE<sp>name<sp>=<sp>'Bolg'" , moria . Goblin . class ) ; moria . Tribe tribe = goblin . getTribe ( ) ; "<AssertPlaceHolder>" ; } catch ( java . sql . SQLException e ) { log . error ( "Fail<sp>caused<sp>by<sp>SQL<sp>exception" , e ) ; org . junit . Assert . fail ( ) ; moria . rollback ( ) ; } finally { moria . close ( ) ; } } getTribe ( ) { return get ( "tribe_id" , moria . Tribe . class ) ; }
org . junit . Assert . assertNotNull ( tribe )
testEmptyStatement ( ) { org . wikidata . wdtk . datamodel . interfaces . ItemIdValue i = org . wikidata . wdtk . datamodel . interfaces . ItemIdValue . NULL ; org . wikidata . wdtk . datamodel . interfaces . PropertyIdValue p = org . wikidata . wdtk . datamodel . interfaces . PropertyIdValue . NULL ; org . wikidata . wdtk . datamodel . interfaces . Statement stmt1 = org . wikidata . wdtk . datamodel . helpers . Datamodel . makeStatement ( i , org . wikidata . wdtk . datamodel . helpers . Datamodel . makeSomeValueSnak ( p ) , java . util . Collections . emptyList ( ) , java . util . Collections . emptyList ( ) , StatementRank . NORMAL , "" ) ; org . wikidata . wdtk . datamodel . interfaces . Statement stmt2 = org . wikidata . wdtk . datamodel . helpers . StatementBuilder . forSubjectAndProperty ( i , p ) . build ( ) ; "<AssertPlaceHolder>" ; } build ( ) { prepareBuild ( ) ; return factory . getPropertyDocument ( ( ( org . wikidata . wdtk . datamodel . interfaces . PropertyIdValue ) ( this . entityIdValue ) ) , this . labels , this . descriptions , this . aliases , getStatementGroups ( ) , this . datatype , this . revisionId ) ; }
org . junit . Assert . assertEquals ( stmt1 , stmt2 )
testAddFactory ( ) { java . lang . String id = "java2" ; flex . messaging . factories . JavaFactory expected = new flex . messaging . factories . JavaFactory ( ) ; broker . addFactory ( "java2" , expected ) ; flex . messaging . factories . JavaFactory actual = ( ( flex . messaging . factories . JavaFactory ) ( broker . getFactory ( id ) ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( expected , actual )
testGetAdapterWithShellAdapter ( ) { java . lang . Object adapter = shell . getAdapter ( org . eclipse . swt . internal . widgets . IShellAdapter . class ) ; "<AssertPlaceHolder>" ; } getAdapter ( org . eclipse . swt . widgets . Shell ) { return shell . getAdapter ( org . eclipse . swt . internal . widgets . IShellAdapter . class ) ; }
org . junit . Assert . assertNotNull ( adapter )