input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testSimplifyOrTerms ( ) { final org . apache . calcite . rel . type . RelDataType intType = typeFactory . createSqlType ( SqlTypeName . INTEGER ) ; final org . apache . calcite . rel . type . RelDataType rowType = typeFactory . builder ( ) . add ( "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 1 , intType ) . nullable ( false ) . add ( "b" , intType ) . nullable ( true ) . add ( "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 7 , intType ) . nullable ( true ) . build ( ) ; final org . apache . calcite . rex . RexDynamicParam range = rexBuilder . makeDynamicParam ( rowType , 0 ) ; final org . apache . calcite . rex . RexNode aRef = rexBuilder . makeFieldAccess ( range , 0 ) ; final org . apache . calcite . rex . RexNode bRef = rexBuilder . makeFieldAccess ( range , 1 ) ; final org . apache . calcite . rex . RexNode cRef = rexBuilder . makeFieldAccess ( range , 2 ) ; final org . apache . calcite . rex . RexLiteral literal1 = rexBuilder . makeExactLiteral ( BigDecimal . ONE ) ; final org . apache . calcite . rex . RexLiteral literal2 = rexBuilder . makeExactLiteral ( java . math . BigDecimal . valueOf ( 2 ) ) ; final org . apache . calcite . rex . RexLiteral literal3 = rexBuilder . makeExactLiteral ( java . math . BigDecimal . valueOf ( 3 ) ) ; final org . apache . calcite . rex . RexLiteral literal4 = rexBuilder . makeExactLiteral ( java . math . BigDecimal . valueOf ( 4 ) ) ; checkSimplifyFilter ( or ( ne ( aRef , literal1 ) , eq ( aRef , literal1 ) ) , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 4 ) ; checkSimplifyFilter ( or ( eq ( aRef , literal1 ) , ne ( aRef , literal1 ) ) , "OR(=(?0.a,<sp>1),<sp><>(?0.a,<sp>1))" ) ; final org . apache . calcite . rex . RexNode neOrEq = or ( ne ( bRef , literal1 ) , eq ( bRef , literal1 ) ) ; checkSimplifyFilter ( neOrEq , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 2 ) ; final org . apache . calcite . rex . RexNode simplified = this . simplify . simplifyUnknownAs ( neOrEq , RexUnknownAs . UNKNOWN ) ; "<AssertPlaceHolder>" ; checkSimplifyFilter ( or ( isNull ( aRef ) , isNotNull ( aRef ) ) , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 4 ) ; checkSimplifyFilter ( or ( isNotNull ( aRef ) , isNull ( aRef ) ) , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 4 ) ; checkSimplifyFilter ( or ( isNotNull ( bRef ) , isNull ( bRef ) ) , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 4 ) ; checkSimplifyFilter ( or ( isNotNull ( bRef ) , isNull ( cRef ) ) , "OR(IS<sp>NOT<sp>NULL(?0.b),<sp>IS<sp>NULL(?0.c))" ) ; checkSimplifyFilter ( or ( isNull ( bRef ) , isNotFalse ( bRef ) ) , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 5 ) ; checkSimplifyFilter ( and ( or ( eq ( bRef , literal1 ) , eq ( bRef , literal2 ) ) , eq ( bRef , literal2 ) , eq ( aRef , literal3 ) , or ( eq ( aRef , literal3 ) , eq ( aRef , literal4 ) ) ) , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 8 ) ; checkSimplify3 ( or ( lt ( vInt ( ) , nullInt ) , ne ( literal ( 0 ) , vInt ( ) ) ) , "OR(null,<sp><>(0,<sp>?0.int0))" , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 3 , "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 4 ) ; } toString ( ) { return args [ 0 ] ; }
org . junit . Assert . assertThat ( simplified . toString ( ) , org . hamcrest . CoreMatchers . equalTo ( "b<sp>!=<sp>1<sp>or<sp>b<sp>=<sp>1" 2 ) )
testAvgAggregation ( ) { com . liferay . portal . search . aggregation . metrics . AvgAggregation avgAggregation = com . liferay . portal . search . aggregations . test . AggregationsInstantiationTest . _aggregations . avg ( "name" , "field" ) ; "<AssertPlaceHolder>" ; } avg ( java . lang . String , java . lang . String ) { return new com . liferay . portal . search . internal . aggregation . metrics . AvgAggregationImpl ( name , field ) ; }
org . junit . Assert . assertNotNull ( avgAggregation )
testEigrpExternalRouteClone ( ) { org . batfish . datamodel . EigrpExternalRoute route = org . batfish . datamodel . EigrpExternalRoute . builder ( ) . setNetwork ( org . batfish . datamodel . Prefix . parse ( "1.1.1.1/32" ) ) . setEigrpMetric ( org . batfish . datamodel . eigrp . EigrpMetric . builder ( ) . setBandwidth ( 1000.0 ) . setDelay ( 2.0 ) . setMode ( EigrpProcessMode . NAMED ) . build ( ) ) . setProcessAsn ( 1L ) . setDestinationAsn ( 2L ) . build ( ) ; "<AssertPlaceHolder>" ; } clone ( java . lang . Object , java . lang . Class ) { return org . batfish . common . util . BatfishObjectMapper . MAPPER . readValue ( org . batfish . common . util . BatfishObjectMapper . WRITER . writeValueAsBytes ( o ) , clazz ) ; }
org . junit . Assert . assertThat ( org . batfish . common . util . BatfishObjectMapper . clone ( route , org . batfish . datamodel . EigrpExternalRoute . class ) , org . hamcrest . Matchers . equalTo ( route ) )
testSayHello ( ) { HelloBean b = new HelloBean ( ) ; "<AssertPlaceHolder>" ; } sayHello ( ) { return "Hello" ; }
org . junit . Assert . assertTrue ( b . sayHello ( ) . equals ( "Hello" ) )
testCastToBigInt ( ) { java . lang . String result = testDialect . getSqlFrom ( new org . alfasoftware . morf . sql . element . Cast ( new org . alfasoftware . morf . sql . element . FieldReference ( "value" ) , org . alfasoftware . morf . metadata . DataType . BIG_INTEGER , 10 ) ) ; "<AssertPlaceHolder>" ; } expectedBigIntCast ( ) { return "CAST(value<sp>AS<sp>BIGINT)" ; }
org . junit . Assert . assertEquals ( expectedBigIntCast ( ) , result )
testMapInhoud1 ( ) { final java . sql . Timestamp timestamp = new java . sql . Timestamp ( 1000L ) ; final nl . bzk . algemeenbrp . dal . domein . brp . entity . AdministratieveHandeling administratieveHandeling = new nl . bzk . algemeenbrp . dal . domein . brp . entity . AdministratieveHandeling ( new nl . bzk . algemeenbrp . dal . domein . brp . entity . Partij ( "partijnaam" , "000042" ) , nl . bzk . algemeenbrp . dal . domein . brp . enums . SoortAdministratieveHandeling . GBA_BIJHOUDING_ACTUEEL , new java . sql . Timestamp ( java . lang . System . currentTimeMillis ( ) ) ) ; final nl . bzk . algemeenbrp . dal . domein . brp . entity . PersoonAfgeleidAdministratiefHistorie historie = new nl . bzk . algemeenbrp . dal . domein . brp . entity . PersoonAfgeleidAdministratiefHistorie ( ( ( short ) ( 1 ) ) , new nl . bzk . algemeenbrp . dal . domein . brp . entity . Persoon ( nl . bzk . algemeenbrp . dal . domein . brp . enums . SoortPersoon . INGESCHREVENE ) , administratieveHandeling , timestamp ) ; final nl . bzk . migratiebrp . conversie . model . brp . groep . BrpPersoonAfgeleidAdministratiefInhoud result = mapper . mapInhoud ( historie , brpOnderzoekMapper ) ; "<AssertPlaceHolder>" ; } mapInhoud ( nl . bzk . algemeenbrp . dal . domein . brp . entity . PersoonAdresHistorie , nl . bzk . migratiebrp . synchronisatie . dal . service . impl . mapper . strategie . BrpOnderzoekMapper ) { final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpSoortAdresCode soortAdresCode = nl . bzk . migratiebrp . synchronisatie . dal . service . impl . mapper . BrpMapperUtil . mapBrpSoortAdresCode ( historie . getSoortAdres ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_SOORTCODE , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpRedenWijzigingVerblijfCode redenWijzigingVerblijfCode = nl . bzk . migratiebrp . synchronisatie . dal . service . impl . mapper . BrpMapperUtil . mapBrpRedenWijzigingVerblijfCode ( historie . getRedenWijziging ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_REDENWIJZIGINGCODE , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpAangeverCode aangeverCode = nl . bzk . migratiebrp . synchronisatie . dal . service . impl . mapper . BrpMapperUtil . mapBrpAangeverCode ( historie . getAangeverAdreshouding ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_AANGEVERADRESHOUDINGCODE , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpDatum datumAanvangAdreshouding = nl . bzk . migratiebrp . synchronisatie . dal . service . impl . mapper . BrpMapperUtil . mapDatum ( historie . getDatumAanvangAdreshouding ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_DATUMAANVANGADRESHOUDING , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString identificatiecodeAdresseerbaarObject = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString . wrap ( historie . getIdentificatiecodeAdresseerbaarObject ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_IDENTIFICATIECODEADRESSEERBAAROBJECT , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString identificatiecodeNummeraanduiding = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString . wrap ( historie . getIdentificatiecodeNummeraanduiding ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_IDENTIFICATIECODENUMMERAANDUIDING , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpGemeenteCode gemeenteCode = nl . bzk . migratiebrp . synchronisatie . dal . service . impl . mapper . BrpMapperUtil . mapBrpGemeenteCode ( historie . getGemeente ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_GEMEENTECODE , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString naamOpenbareRuimte = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString . wrap ( historie . getNaamOpenbareRuimte ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_NAAMOPENBARERUIMTE , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString afgekorteNaamOpenbareRuimte = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString . wrap ( historie . getAfgekorteNaamOpenbareRuimte ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_AFGEKORTENAAMOPENBARERUIMTE , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString gemeentedeel = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString . wrap ( historie . getGemeentedeel ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_GEMEENTEDEEL , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpInteger huisnummer = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpInteger . wrap ( historie . getHuisnummer ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_HUISNUMMER , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpCharacter huisletter = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpCharacter . wrap ( historie . getHuisletter ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie , Element . PERSOON_ADRES_HUISLETTER , true ) ) ; final nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString huisnummerToevoeging = nl . bzk . migratiebrp . conversie . model . brp . attribuut . BrpString . wrap ( historie . getHuisnummertoevoeging ( ) , brpOnderzoekMapper . bepaalOnderzoek ( historie ,
org . junit . Assert . assertNotNull ( result )
TestComplexEmptyGet ( ) { java . util . List r = com . groupon . mapreduce . mongo . JobUtil . get ( m , "d.y.foo" ) ; "<AssertPlaceHolder>" ; } get ( java . util . Map , java . lang . String ) { java . lang . String [ ] fields = path . split ( "\\." ) ; java . util . List found = new java . util . ArrayList ( ) ; com . groupon . mapreduce . mongo . JobUtil . get ( map , fields , 0 , found ) ; return found ; }
org . junit . Assert . assertEquals ( r . size ( ) , 0 )
getAllTypeLibrariesNames ( ) { java . util . Set < java . lang . String > set = org . ebayopensource . turmeric . tools . library . TypeLibraryRegistryInterfaceTest . m_soaTypeRegistry . getAllTypeLibrariesNames ( ) ; "<AssertPlaceHolder>" ; } getAllTypeLibrariesNames ( ) { java . util . Set < java . lang . String > set = org . ebayopensource . turmeric . tools . library . TypeLibraryRegistryInterfaceTest . m_soaTypeRegistry . getAllTypeLibrariesNames ( ) ; org . junit . Assert . assertTrue ( set . contains ( org . ebayopensource . turmeric . tools . library . TypeLibraryRegistryInterfaceTest . m_libraryName ) ) ; }
org . junit . Assert . assertTrue ( set . contains ( org . ebayopensource . turmeric . tools . library . TypeLibraryRegistryInterfaceTest . m_libraryName ) )
test1 ( ) { java . util . Set < io . djigger . monitoring . java . instrumentation . InstrumentSubscription > subscription = new java . util . HashSet < io . djigger . monitoring . java . instrumentation . InstrumentSubscription > ( ) ; io . djigger . monitoring . java . model . StackTraceElement el = new io . djigger . monitoring . java . model . StackTraceElement ( "Class1" , "Method1" , null , ( - 1 ) ) ; io . djigger . monitoring . java . model . StackTraceElement [ ] els = new io . djigger . monitoring . java . model . StackTraceElement [ ] { el } ; subscription . add ( new io . djigger . monitoring . java . instrumentation . subscription . RealNodePathSubscription ( els , false ) ) ; io . djigger . monitoring . java . model . StackTraceElement el2 = new io . djigger . monitoring . java . model . StackTraceElement ( "Class1" , "Method1" , null , ( - 1 ) ) ; io . djigger . monitoring . java . model . StackTraceElement [ ] els2 = new io . djigger . monitoring . java . model . StackTraceElement [ ] { el2 } ; subscription . remove ( new io . djigger . monitoring . java . instrumentation . subscription . RealNodePathSubscription ( els2 , false ) ) ; "<AssertPlaceHolder>" ; } add ( long , io . djigger . monitoring . java . model . ThreadInfo ) { int rangeId = rangeDefinition . getRangeId ( time ) ; rangeTable [ rangeId ] . add ( thread ) ; }
org . junit . Assert . assertEquals ( 0 , subscription . size ( ) )
testMatches ( ) { final org . osgi . framework . Filter filter = context . mock ( org . osgi . framework . Filter . class ) ; event . setReadOnly ( true ) ; final java . util . Dictionary < java . lang . String , java . lang . Object > properties = event . getProperties ( ) ; context . checking ( new org . jmock . Expectations ( ) { { oneOf ( filter ) . match ( properties ) ; will ( returnValue ( true ) ) ; } } ) ; "<AssertPlaceHolder>" ; } matches ( org . osgi . framework . Filter ) { return filter . match ( getProperties ( ) ) ; }
org . junit . Assert . assertTrue ( event . matches ( filter ) )
distcpShouldNotBeCalledOnlyForOneFile ( ) { mockStatic ( org . apache . hadoop . security . UserGroupInformation . class ) ; when ( org . apache . hadoop . security . UserGroupInformation . getCurrentUser ( ) ) . thenReturn ( mock ( org . apache . hadoop . security . UserGroupInformation . class ) ) ; org . apache . hadoop . hive . conf . HiveConf conf = org . mockito . Mockito . spy ( new org . apache . hadoop . hive . conf . HiveConf ( ) ) ; doReturn ( 1L ) . when ( conf ) . getLong ( HiveConf . ConfVars . HIVE_EXEC_COPYFILE_MAXSIZE . varname , ( ( 32L * 1024 ) * 1024 ) ) ; org . apache . hadoop . hive . ql . parse . repl . CopyUtils copyUtils = new org . apache . hadoop . hive . ql . parse . repl . CopyUtils ( "" , conf , null ) ; long MB_128 = ( 128 * 1024 ) * 1024 ; "<AssertPlaceHolder>" ; } limitReachedForLocalCopy ( long , long ) { boolean result = ( size > ( maxCopyFileSize ) ) && ( numberOfFiles > ( maxNumberOfFiles ) ) ; if ( result ) { org . apache . hadoop . hive . ql . parse . repl . CopyUtils . LOG . info ( "Source<sp>is<sp>{}<sp>bytes.<sp>(MAX:<sp>{})" , size , maxCopyFileSize ) ; org . apache . hadoop . hive . ql . parse . repl . CopyUtils . LOG . info ( "Source<sp>is<sp>{}<sp>files.<sp>(MAX:<sp>{})" , numberOfFiles , maxNumberOfFiles ) ; org . apache . hadoop . hive . ql . parse . repl . CopyUtils . LOG . info ( "going<sp>to<sp>launch<sp>distributed<sp>copy<sp>(distcp)<sp>job." ) ; } return result ; }
org . junit . Assert . assertFalse ( copyUtils . limitReachedForLocalCopy ( MB_128 , 1L ) )
fixPathCommonTest ( ) { java . nio . file . Path inputPath = java . nio . file . Paths . get ( "subdir/subdir2/foo/bar/x.file" ) ; java . nio . file . Path expecting = java . nio . file . Paths . get ( "subdir/subdir2/foo/bar/x.file" ) ; "<AssertPlaceHolder>" ; } fixPath ( java . nio . file . Path ) { int i = 0 ; while ( i < ( p . getNameCount ( ) ) ) { if ( p . getName ( i ) . toString ( ) . equals ( "." ) ) { p = com . facebook . buck . io . file . MorePaths . dropPathPart ( p , i ) ; } else { i ++ ; } } return p ; }
org . junit . Assert . assertEquals ( expecting , com . facebook . buck . io . file . MorePaths . fixPath ( inputPath ) )
shouldCacheAllAuthors ( ) { int first = - 1 ; int second = - 1 ; org . apache . ibatis . session . SqlSession session = org . apache . ibatis . session . SqlSessionTest . sqlMapper . openSession ( ) ; try { java . util . List < org . apache . ibatis . domain . blog . Author > authors = session . selectList ( "com.domain.CachedAuthorMapper.selectAllAuthors" ) ; first = java . lang . System . identityHashCode ( authors ) ; session . commit ( ) ; } finally { session . close ( ) ; } session = org . apache . ibatis . session . SqlSessionTest . sqlMapper . openSession ( ) ; try { java . util . List < org . apache . ibatis . domain . blog . Author > authors = session . selectList ( "com.domain.CachedAuthorMapper.selectAllAuthors" ) ; second = java . lang . System . identityHashCode ( authors ) ; } finally { session . close ( ) ; } "<AssertPlaceHolder>" ; } close ( ) { try { executor . close ( isCommitOrRollbackRequired ( false ) ) ; closeCursors ( ) ; dirty = false ; } finally { org . apache . ibatis . executor . ErrorContext . instance ( ) . reset ( ) ; } }
org . junit . Assert . assertEquals ( first , second )
testUtilitySerialization ( ) { final org . apache . hadoop . fs . Path p = new org . apache . hadoop . fs . Path ( "hdfs://yaks:4344/dingos/f" ) ; org . apache . hadoop . fs . FileStatus stat = new org . apache . hadoop . fs . FileStatus ( 1024L , false , 3 , ( 1L << 31 ) , 12345678L , 87654321L , org . apache . hadoop . fs . permission . FsPermission . createImmutable ( ( ( short ) ( 73 ) ) ) , "hadoop" , "unqbbc" , null , p ) ; org . apache . hadoop . fs . protocolPB . FileStatusProto fsp = org . apache . hadoop . fs . protocolPB . PBHelper . convert ( stat ) ; org . apache . hadoop . fs . FileStatus stat2 = org . apache . hadoop . fs . protocolPB . PBHelper . convert ( fsp ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . fs . protocolPB . TestFSSerialization . checkFields ( stat , stat2 ) ; } convert ( org . apache . hadoop . fs . protocolPB . FsPermissionProto ) { return new org . apache . hadoop . fs . permission . FsPermission ( ( ( short ) ( proto . getPerm ( ) ) ) ) ; }
org . junit . Assert . assertEquals ( stat , stat2 )
testTaskRequestWithFutureStartMsDoesNotGetRun ( ) { org . apache . kafka . common . utils . MockTime time = new org . apache . kafka . common . utils . MockTime ( 0 , 0 , 0 ) ; org . apache . kafka . common . utils . Scheduler scheduler = new org . apache . kafka . common . utils . MockScheduler ( time ) ; try ( org . apache . kafka . trogdor . common . MiniTrogdorCluster cluster = new org . apache . kafka . trogdor . common . MiniTrogdorCluster . Builder ( ) . addCoordinator ( "node01" ) . addAgent ( "node02" ) . scheduler ( scheduler ) . build ( ) ) { org . apache . kafka . trogdor . task . NoOpTaskSpec fooSpec = new org . apache . kafka . trogdor . task . NoOpTaskSpec ( 1000 , 500 ) ; time . sleep ( 999 ) ; org . apache . kafka . trogdor . coordinator . CoordinatorClient coordinatorClient = cluster . coordinatorClient ( ) ; coordinatorClient . createTask ( new org . apache . kafka . trogdor . rest . CreateTaskRequest ( "fooSpec" , fooSpec ) ) ; org . apache . kafka . trogdor . rest . TaskState expectedState = new org . apache . kafka . trogdor . common . ExpectedTasks . ExpectedTaskBuilder ( "fooSpec" ) . taskState ( new org . apache . kafka . trogdor . rest . TaskPending ( fooSpec ) ) . build ( ) . taskState ( ) ; org . apache . kafka . trogdor . rest . TaskState resp = coordinatorClient . task ( new org . apache . kafka . trogdor . rest . TaskRequest ( "fooSpec" ) ) ; "<AssertPlaceHolder>" ; } } task ( org . apache . kafka . trogdor . rest . TaskRequest ) { java . lang . String uri = javax . ws . rs . core . UriBuilder . fromPath ( url ( "/coordinator/tasks/{taskId}" ) ) . build ( request . taskId ( ) ) . toString ( ) ; org . apache . kafka . trogdor . rest . JsonRestServer . HttpResponse < org . apache . kafka . trogdor . rest . TaskState > resp = org . apache . kafka . trogdor . rest . JsonRestServer . httpRequest ( log , uri , "GET" , null , new com . fasterxml . jackson . core . type . TypeReference < org . apache . kafka . trogdor . rest . TaskState > ( ) { } , maxTries ) ; return resp . body ( ) ; }
org . junit . Assert . assertEquals ( expectedState , resp )
testUnwrapDecryptNull ( ) { final java . lang . String wrappedEncryptedValue = null ; ddf . security . encryption . impl . EncryptionServiceImplTest . LOGGER . debug ( "Original<sp>wrapped<sp>encrypted<sp>value<sp>is:<sp>null" ) ; final ddf . security . encryption . impl . EncryptionServiceImpl encryptionService = new ddf . security . encryption . impl . EncryptionServiceImpl ( ) ; final java . lang . String decryptedValue = encryptionService . decryptValue ( wrappedEncryptedValue ) ; "<AssertPlaceHolder>" ; } decryptValue ( java . lang . String ) { if ( org . apache . commons . lang3 . StringUtils . isEmpty ( wrappedEncryptedValue ) ) { return wrappedEncryptedValue ; } java . lang . String encryptedValue = unwrapEncryptedValue ( wrappedEncryptedValue ) ; if ( wrappedEncryptedValue . equals ( encryptedValue ) ) { return wrappedEncryptedValue ; } return decrypt ( encryptedValue ) ; }
org . junit . Assert . assertNull ( decryptedValue )
execute_emptypage ( ) { org . mitre . data . AbstractPageOperationTemplateTest . CountingPageOperation op = new org . mitre . data . AbstractPageOperationTemplateTest . EmptyPageCountingPageOperation ( Integer . MAX_VALUE , Long . MAX_VALUE ) ; op . execute ( ) ; "<AssertPlaceHolder>" ; } getCounter ( ) { return counter ; }
org . junit . Assert . assertEquals ( 0L , op . getCounter ( ) )
testSufficientConfig ( ) { org . dcache . gplazma . strategies . SessionStrategy strategy = strategyFactory . newSessionStrategy ( ) ; "<AssertPlaceHolder>" ; strategy . setPlugins ( successSufficientPlugins ) ; java . util . Set < java . security . Principal > authorizedPrincipals = com . google . common . collect . Sets . newHashSet ( ) ; java . util . Set < java . lang . Object > sessionAttributes = com . google . common . collect . Sets . newHashSet ( ) ; strategy . session ( org . dcache . gplazma . strategies . SessionStrategyTests . IGNORING_LOGIN_MONITOR , authorizedPrincipals , sessionAttributes ) ; } newSessionStrategy ( ) { return new org . dcache . gplazma . strategies . DefaultSessionStrategy ( ) ; }
org . junit . Assert . assertNotNull ( strategy )
testToString ( ) { java . lang . String principal = "my_special_principal" ; org . apache . accumulo . core . clientImpl . AuthenticationTokenIdentifier token = new org . apache . accumulo . core . clientImpl . AuthenticationTokenIdentifier ( principal ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( this . key ) + "<sp>-<sp>" ) + ( description ) ; }
org . junit . Assert . assertTrue ( token . toString ( ) . contains ( principal ) )
testCreateSequentialRecurrence1 ( ) { "<AssertPlaceHolder>" ; } createSequentialRecurrence ( org . pentaho . platform . plugin . services . importexport . exportManifest . bindings . SequentialRecurrence ) { return new javax . xml . bind . JAXBElement < org . pentaho . platform . plugin . services . importexport . exportManifest . bindings . SequentialRecurrence > ( org . pentaho . platform . plugin . services . importexport . exportManifest . bindings . ObjectFactory . _SequentialRecurrence_QNAME , org . pentaho . platform . plugin . services . importexport . exportManifest . bindings . SequentialRecurrence . class , null , value ) ; }
org . junit . Assert . assertNotNull ( factory . createSequentialRecurrence ( mock ( org . pentaho . platform . plugin . services . importexport . exportManifest . bindings . SequentialRecurrence . class ) ) )
testBuildWithParametersWithOrderBy ( ) { java . lang . String abbrName = "AN" ; java . lang . String name = "fdsfds" ; org . lnu . is . domain . employee . type . EmployeeType context = new org . lnu . is . domain . employee . type . EmployeeType ( ) ; context . setAbbrName ( abbrName ) ; context . setName ( name ) ; org . lnu . is . pagination . OrderBy orderBy1 = new org . lnu . is . pagination . OrderBy ( "abbrName" , org . lnu . is . pagination . OrderByType . ASC ) ; org . lnu . is . pagination . OrderBy orderBy2 = new org . lnu . is . pagination . OrderBy ( "name" , org . lnu . is . pagination . OrderByType . DESC ) ; java . util . List < org . lnu . is . pagination . OrderBy > orders = java . util . Arrays . asList ( orderBy1 , orderBy2 ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>EmployeeType<sp>e<sp>WHERE<sp>(<sp>e.name<sp>LIKE<sp>CONCAT('%',:name,'%')<sp>AND<sp>e.abbrName<sp>LIKE<sp>CONCAT('%',:abbrName,'%')<sp>)<sp>AND<sp>e.status=:status<sp>AND<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>ORDER<sp>BY<sp>e.abbrName<sp>ASC,<sp>e.name<sp>DESC" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . employee . type . EmployeeType > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; pagedSearch . setOrders ( orders ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setOrders ( java . util . List ) { this . orders = orders ; }
org . junit . Assert . assertEquals ( expectedQuery , actualQuery )
emptyZeroCharBuffersize ( ) { final javax . json . JsonReader reader = javax . json . Json . createReaderFactory ( new java . util . HashMap < java . lang . String , java . lang . Object > ( ) { { put ( "org.apache.johnzon.default-char-buffer" , "0" ) ; } } ) . createReader ( java . lang . Thread . currentThread ( ) . getContextClassLoader ( ) . getResourceAsStream ( "json/empty.json" ) , org . apache . johnzon . core . JsonReaderImplTest . utf8Charset ) ; "<AssertPlaceHolder>" ; reader . readObject ( ) ; reader . close ( ) ; } createReader ( java . io . InputStream , java . nio . charset . Charset ) { return new org . apache . johnzon . core . JsonReaderImpl ( parserFactory . createInternalParser ( in , charset ) ) ; }
org . junit . Assert . assertNotNull ( reader )
testReadComponentType ( ) { java . io . InputStream is = getClass ( ) . getResourceAsStream ( "CalculatorImpl.componentType" ) ; javax . xml . stream . XMLStreamReader reader = org . apache . tuscany . sca . binding . ws . xml . ReadTestCase . inputFactory . createXMLStreamReader ( is ) ; org . apache . tuscany . sca . assembly . ComponentType componentType = ( ( org . apache . tuscany . sca . assembly . ComponentType ) ( org . apache . tuscany . sca . binding . ws . xml . ReadTestCase . staxProcessor . read ( reader , org . apache . tuscany . sca . binding . ws . xml . ReadTestCase . context ) ) ) ; "<AssertPlaceHolder>" ; } read ( javax . xml . stream . XMLStreamReader , org . apache . tuscany . sca . contribution . processor . ProcessorContext ) { org . apache . tuscany . sca . monitor . Monitor monitor = context . getMonitor ( ) ; org . apache . tuscany . sca . binding . ws . WebServiceBinding wsBinding = wsFactory . createWebServiceBinding ( ) ; wsBinding . setUnresolved ( true ) ; policyProcessor . readPolicies ( wsBinding , reader ) ; java . lang . String name = reader . getAttributeValue ( null , org . apache . tuscany . sca . binding . ws . xml . NAME ) ; if ( name != null ) { wsBinding . setName ( name ) ; } java . util . List < java . lang . String > endpointSpecifications = new java . util . ArrayList < java . lang . String > ( ) ; java . lang . String uri = getURIString ( reader , java . net . URI ) ; if ( uri != null ) { wsBinding . setURI ( uri ) ; wsBinding . setUserSpecifiedURI ( uri ) ; if ( ( context . getParentModel ( ) ) instanceof org . apache . tuscany . sca . assembly . Reference ) { try { java . net . URI tmpURI = new java . net . URI ( uri ) ; if ( ! ( tmpURI . isAbsolute ( ) ) ) { error ( monitor , "URINotAbsolute" , reader , uri ) ; } } catch ( java . net . URISyntaxException ex ) { error ( monitor , "InvalidWsdlElementAttr" 0 , reader , ex . getMessage ( ) ) ; } endpointSpecifications . add ( "InvalidWsdlElementAttr" 9 ) ; } if ( ( ( context . getParentModel ( ) ) instanceof org . apache . tuscany . sca . assembly . Callback ) && ( ( ( ( org . apache . tuscany . sca . assembly . Callback ) ( context . getParentModel ( ) ) ) . getParentContract ( ) ) instanceof org . apache . tuscany . sca . assembly . Service ) ) { error ( monitor , "InvalidWsdlElementAttr" 8 , reader , uri ) ; } } java . lang . Boolean wsdlElementIsBinding = null ; java . lang . String wsdlElement = getURIString ( reader , org . apache . tuscany . sca . binding . ws . xml . WSDL_ELEMENT ) ; if ( wsdlElement != null ) { int index = wsdlElement . indexOf ( '#' ) ; if ( index == ( - 1 ) ) { error ( monitor , "InvalidWsdlElementAttr" , reader , wsdlElement ) ; return wsBinding ; } java . lang . String namespace = wsdlElement . substring ( 0 , index ) ; wsBinding . setNamespace ( namespace ) ; java . lang . String localName = wsdlElement . substring ( ( index + 1 ) ) ; if ( localName . startsWith ( "InvalidWsdlElementAttr" 6 ) ) { if ( ( context . getParentModel ( ) ) instanceof org . apache . tuscany . sca . assembly . Service ) { error ( monitor , "InvalidWsdlElementAttr" 4 , reader , wsdlElement ) ; } localName = localName . substring ( "wsdl.service(" . length ( ) , ( ( localName . length ( ) ) - 1 ) ) ; wsBinding . setServiceName ( new javax . xml . namespace . QName ( namespace , localName ) ) ; endpointSpecifications . add ( "#wsdl.service" ) ; } else if ( localName . startsWith ( "InvalidWsdlElementAttr" 3 ) ) { localName = localName . substring ( "wsdl.port(" . length ( ) , ( ( localName . length ( ) ) - 1 ) ) ; int s = localName . indexOf ( '/' ) ; if ( s == ( - 1 ) ) { error ( monitor , "InvalidWsdlElementAttr" , reader , wsdlElement ) ; } else { wsBinding . setServiceName ( new javax . xml . namespace . QName ( namespace , localName . substring ( 0 , s ) ) ) ; wsBinding . setPortName ( localName . substring ( ( s + 1 ) ) ) ; } endpointSpecifications . add ( "InvalidWsdlElementAttr" 7 ) ; } else if ( localName . startsWith ( "wsdl.endpoint" ) ) { localName = localName . substring ( "wsdl.endpoint(" . length ( ) , ( ( localName . length ( ) ) - 1 ) ) ; int s = localName . indexOf ( '/' ) ; if ( s == ( - 1 ) ) { error ( monitor , "InvalidWsdlElementAttr" , reader , wsdlElement ) ; } else { wsBinding . setServiceName ( new javax . xml . namespace . QName ( namespace , localName . substring ( 0 , s ) ) ) ; wsBinding . setEndpointName ( localName . substring ( ( s + 1 ) ) ) ; } } else if ( localName . startsWith ( "URINotAbsolute" 4 ) ) { localName = localName . substring ( "wsdl.binding(" . length ( ) , ( ( localName . length ( ) ) - 1 ) ) ; wsBinding
org . junit . Assert . assertNotNull ( componentType )
testClientNoArgs ( ) { int rc = newClient ( ) . start ( new java . lang . String [ ] { } ) ; "<AssertPlaceHolder>" ; } start ( java . lang . String [ ] ) { com . github . perlundq . yajsync . internal . util . ArgumentParser argsParser = com . github . perlundq . yajsync . internal . util . ArgumentParser . newNoUnnamed ( getClass ( ) . getSimpleName ( ) ) ; try { argsParser . addHelpTextDestination ( _out ) ; for ( com . github . perlundq . yajsync . internal . util . Option o : options ( ) ) { argsParser . add ( o ) ; } for ( com . github . perlundq . yajsync . internal . util . Option o : _moduleProvider . options ( ) ) { argsParser . add ( o ) ; } com . github . perlundq . yajsync . internal . util . ArgumentParser . Status rc = argsParser . parse ( java . util . Arrays . asList ( args ) ) ; if ( rc != ( ArgumentParser . Status . CONTINUE ) ) { return rc == ( ArgumentParser . Status . EXIT_OK ) ? 0 : 1 ; } } catch ( com . github . perlundq . yajsync . internal . util . ArgumentParsingError e ) { _err . println ( e . getMessage ( ) ) ; _err . println ( argsParser . toUsageString ( ) ) ; return - 1 ; } java . util . logging . Level logLevel = com . github . perlundq . yajsync . internal . util . Util . getLogLevelForNumber ( ( ( com . github . perlundq . yajsync . internal . util . Util . WARNING_LOG_LEVEL_NUM ) + ( _verbosity ) ) ) ; com . github . perlundq . yajsync . internal . util . Util . setRootLogLevel ( logLevel ) ; com . github . perlundq . yajsync . net . ServerChannelFactory socketFactory = ( _isTLS ) ? new com . github . perlundq . yajsync . net . SSLServerChannelFactory ( ) . setWantClientAuth ( true ) : new com . github . perlundq . yajsync . net . StandardServerChannelFactory ( ) ; socketFactory . setReuseAddress ( true ) ; boolean isInterruptible = ! ( _isTLS ) ; java . util . concurrent . ExecutorService executor = java . util . concurrent . Executors . newFixedThreadPool ( _numThreads ) ; com . github . perlundq . yajsync . RsyncServer server = _serverBuilder . build ( executor ) ; try ( com . github . perlundq . yajsync . net . ServerChannel listenSock = socketFactory . open ( _address , _port , _timeout ) ) { if ( ( _isListeningLatch ) != null ) { _isListeningLatch . countDown ( ) ; } while ( true ) { com . github . perlundq . yajsync . net . DuplexByteChannel sock = listenSock . accept ( ) ; java . util . concurrent . Callable < java . lang . Boolean > c = createCallable ( server , sock , isInterruptible ) ; executor . submit ( c ) ; } } finally { if ( com . github . perlundq . yajsync . ui . YajsyncServer . _log . isLoggable ( Level . INFO ) ) { com . github . perlundq . yajsync . ui . YajsyncServer . _log . info ( "shutting<sp>down..." ) ; } executor . shutdown ( ) ; _moduleProvider . close ( ) ; while ( ! ( executor . awaitTermination ( 5 , TimeUnit . MINUTES ) ) ) { com . github . perlundq . yajsync . ui . YajsyncServer . _log . info ( ( "some<sp>sessions<sp>are<sp>still<sp>running,<sp>waiting<sp>for<sp>them<sp>" + "to<sp>finish<sp>before<sp>exiting" ) ) ; } if ( com . github . perlundq . yajsync . ui . YajsyncServer . _log . isLoggable ( Level . INFO ) ) { com . github . perlundq . yajsync . ui . YajsyncServer . _log . info ( "done" ) ; } } }
org . junit . Assert . assertTrue ( ( rc == ( - 1 ) ) )
testUnixModeInAddRaw ( ) { final java . io . File [ ] tmp = createTempDirAndFile ( ) ; final java . io . File a1 = java . io . File . createTempFile ( "unixModeBits." , ".zip" , tmp [ 0 ] ) ; try ( final org . apache . commons . compress . archivers . zip . ZipArchiveOutputStream zos = new org . apache . commons . compress . archivers . zip . ZipArchiveOutputStream ( a1 ) ) { final org . apache . commons . compress . archivers . zip . ZipArchiveEntry archiveEntry = new org . apache . commons . compress . archivers . zip . ZipArchiveEntry ( "fred" ) ; archiveEntry . setUnixMode ( 436 ) ; archiveEntry . setMethod ( ZipEntry . DEFLATED ) ; zos . addRawArchiveEntry ( archiveEntry , new java . io . ByteArrayInputStream ( "fud" . getBytes ( ) ) ) ; } try ( final org . apache . commons . compress . archivers . zip . ZipFile zf1 = new org . apache . commons . compress . archivers . zip . ZipFile ( a1 ) ) { final org . apache . commons . compress . archivers . zip . ZipArchiveEntry fred = zf1 . getEntry ( "fred" ) ; "<AssertPlaceHolder>" ; } } getUnixMode ( ) { return isHostOsUnix ( ) ? getMode ( ) : 0 ; }
org . junit . Assert . assertEquals ( 436 , fred . getUnixMode ( ) )
emptyTexts ( ) { java . lang . String diff = com . github . ferstl . maven . pomenforcers . util . SideBySideDiffUtil . diff ( java . util . Collections . emptyList ( ) , java . util . Collections . emptyList ( ) , "" , "" ) ; "<AssertPlaceHolder>" ; } diff ( java . util . Collection , java . util . Collection , java . lang . String , java . lang . String ) { com . github . ferstl . maven . pomenforcers . util . SideBySideDiffUtil . SideBySideContext context = new com . github . ferstl . maven . pomenforcers . util . SideBySideDiffUtil . SideBySideContext ( actual , required , leftTitle , rightTitle ) ; int offset = 0 ; for ( difflib . Delta < java . lang . String > delta : context . deltas ) { difflib . Chunk < java . lang . String > original = delta . getOriginal ( ) ; difflib . Chunk < java . lang . String > revised = delta . getRevised ( ) ; int currentPosition = ( original . getPosition ( ) ) + offset ; switch ( delta . getType ( ) ) { case INSERT : offset += context . expand ( currentPosition , revised . size ( ) ) ; context . setRightContent ( currentPosition , revised . getLines ( ) ) ; break ; case CHANGE : int difference = ( revised . size ( ) ) - ( original . size ( ) ) ; if ( difference > 0 ) { offset += context . expand ( ( currentPosition + ( original . size ( ) ) ) , difference ) ; } else { context . clearRightContent ( ( currentPosition + ( revised . size ( ) ) ) , java . lang . Math . abs ( difference ) ) ; } context . setLeftContent ( currentPosition , original . getLines ( ) ) ; context . setRightContent ( currentPosition , revised . getLines ( ) ) ; break ; case DELETE : context . setLeftContent ( currentPosition , original . getLines ( ) ) ; context . clearRightContent ( currentPosition , original . size ( ) ) ; break ; default : throw new java . lang . IllegalStateException ( ( "Unsupported<sp>delta<sp>type:<sp>" + ( delta . getType ( ) ) ) ) ; } } return context . toString ( ) ; }
org . junit . Assert . assertEquals ( "" , diff )
test ( ) { com . mysema . rdfbean . model . BID bid = new com . mysema . rdfbean . model . BID ( "c" ) ; com . mysema . rdfbean . model . UID uid = new com . mysema . rdfbean . model . UID ( "b:b" ) ; com . mysema . rdfbean . model . LIT lit = new com . mysema . rdfbean . model . LIT ( "a" ) ; com . mysema . rdfbean . model . STMT stmt1 = new com . mysema . rdfbean . model . STMT ( bid , uid , bid ) ; com . mysema . rdfbean . model . STMT stmt2 = new com . mysema . rdfbean . model . STMT ( bid , uid , uid ) ; com . mysema . rdfbean . model . STMT stmt3 = new com . mysema . rdfbean . model . STMT ( bid , uid , lit ) ; com . mysema . rdfbean . model . STMT stmt4 = new com . mysema . rdfbean . model . STMT ( uid , uid , bid ) ; com . mysema . rdfbean . model . STMT stmt5 = new com . mysema . rdfbean . model . STMT ( uid , uid , uid ) ; com . mysema . rdfbean . model . STMT stmt6 = new com . mysema . rdfbean . model . STMT ( uid , uid , lit ) ; java . util . List < com . mysema . rdfbean . model . STMT > stmts = java . util . Arrays . < com . mysema . rdfbean . model . STMT > asList ( stmt6 , stmt5 , stmt4 , stmt3 , stmt2 , stmt1 ) ; java . util . Collections . sort ( stmts , comparator ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( java . util . Arrays . asList ( stmt1 , stmt2 , stmt3 , stmt4 , stmt5 , stmt6 ) , stmts )
addLastLenTest ( ) { org . krakenapps . pcap . util . Buffer buffer = initBuffer4 ( ) ; org . krakenapps . pcap . util . Buffer buffer4 = new org . krakenapps . pcap . util . ChainBuffer ( ) ; byte [ ] testArray13 = new byte [ ] { 71 , 72 , 73 , 74 , 75 , 76 } ; byte [ ] testArray14 = new byte [ ] { 77 , 78 , 79 } ; byte [ ] testArray15 = new byte [ ] { 80 , 81 } ; byte [ ] testArray16 = new byte [ ] { 82 , 83 , 84 , 85 , 86 , 87 , 88 , 89 } ; buffer4 . addLast ( testArray13 ) ; buffer4 . addLast ( testArray14 ) ; buffer4 . addLast ( testArray15 ) ; buffer4 . addLast ( testArray16 ) ; buffer . addLast ( buffer4 , 5 ) ; byte [ ] skip = new byte [ 41 ] ; buffer . gets ( skip ) ; int expected = 71 ; int i = 0 ; while ( i < 5 ) { "<AssertPlaceHolder>" ; expected ++ ; i ++ ; } } get ( ) { return value ; }
org . junit . Assert . assertEquals ( expected , buffer . get ( ) )
testPartialUpdateXMLDscriptorTranc ( ) { System . out . println ( "Running<sp>testPartialUpdateXMLDescriptorTranc" ) ; java . lang . String [ ] filenames = new java . lang . String [ ] { "constraint1.xml" , "constraint1.xml" 5 , "constraint1.xml" 1 , "constraint1.xml" 7 , "constraint5.xml" } ; com . marklogic . client . DatabaseClientFactory . SecurityContext secContext = new com . marklogic . client . DatabaseClientFactory . DigestAuthContext ( "constraint1.xml" 2 , "x" ) ; com . marklogic . client . DatabaseClient client = com . marklogic . client . DatabaseClientFactory . newClient ( com . marklogic . client . functionaltest . TestPartialUpdate . appServerHostname , com . marklogic . client . functionaltest . TestPartialUpdate . uberPort , com . marklogic . client . functionaltest . TestPartialUpdate . dbName , secContext , getConnType ( ) ) ; for ( java . lang . String filename : filenames ) { writeDocumentUsingInputStreamHandle ( client , filename , "/partial-update/" , "XML" ) ; } java . lang . String docId = "/partial-update/constraint1.xml" ; com . marklogic . client . document . XMLDocumentManager docMgr = client . newXMLDocumentManager ( ) ; com . marklogic . client . document . DocumentUriTemplate template = docMgr . newDocumentUriTemplate ( "constraint1.xml" 3 ) ; template . withDirectory ( docId ) ; com . marklogic . client . document . DocumentDescriptor desc = docMgr . newDescriptor ( template . getDirectory ( ) ) ; com . marklogic . client . document . DocumentPatchBuilder patchBldr = docMgr . newPatchBuilder ( ) ; patchBldr . insertFragment ( "constraint1.xml" 0 , Position . LAST_CHILD , "<modified>2013-03-21</modified>" ) ; com . marklogic . client . io . marker . DocumentPatchHandle patchHandle = patchBldr . build ( ) ; com . marklogic . client . Transaction t = client . openTransaction ( "Tranc" ) ; docMgr . patch ( desc , patchHandle , t ) ; t . commit ( ) ; waitForPropertyPropagate ( ) ; java . lang . String content = docMgr . read ( docId , new com . marklogic . client . io . StringHandle ( ) ) . get ( ) ; System . out . println ( ( "After" + content ) ) ; "<AssertPlaceHolder>" ; client . release ( ) ; } get ( ) { return content ; }
org . junit . Assert . assertTrue ( "constraint1.xml" 6 , content . contains ( "constraint1.xml" 4 ) )
when_setName_thenReturnsName ( ) { com . hazelcast . jet . config . JobConfig config = new com . hazelcast . jet . config . JobConfig ( ) ; java . lang . String name = "myJobName" ; config . setName ( name ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; }
org . junit . Assert . assertEquals ( name , config . getName ( ) )
backOffFactorIsConfigurable ( ) { connPerRoute . setMaxPerRoute ( route , 10 ) ; impl . setBackoffFactor ( 0.9 ) ; impl . backOff ( route ) ; "<AssertPlaceHolder>" ; } getMaxPerRoute ( org . apache . hc . client5 . http . HttpRoute ) { return pool . getMaxPerRoute ( route ) ; }
org . junit . Assert . assertEquals ( 9 , connPerRoute . getMaxPerRoute ( route ) )
injectionContextWithMathingKey ( ) { java . lang . String expected = "hello<sp>duke" ; java . util . Map < java . lang . String , java . lang . Object > injectionContext = new java . util . HashMap ( ) ; injectionContext . put ( "name" , expected ) ; com . airhacks . afterburner . injection . PresenterWithField withField = com . airhacks . afterburner . injection . Injector . instantiatePresenter ( com . airhacks . afterburner . injection . PresenterWithField . class , injectionContext :: get ) ; "<AssertPlaceHolder>" ; com . airhacks . afterburner . injection . Injector . forgetAll ( ) ; } getName ( ) { return name ; }
org . junit . Assert . assertThat ( withField . getName ( ) , org . hamcrest . CoreMatchers . is ( expected ) )
testInsertRowInEmptyMatrix ( ) { org . la4j . matrix . Matrix a = m ( org . la4j . matrix . MatrixTest . a ( 1.0 , 0.0 , 0.0 ) , org . la4j . matrix . MatrixTest . a ( 0.0 , 1.0 , 0.0 ) , org . la4j . matrix . MatrixTest . a ( 0.0 , 0.0 , 1.0 ) ) ; org . la4j . matrix . Vector v1 = org . la4j . matrix . MatrixTest . v ( 1.0 , 0.0 , 0.0 ) ; org . la4j . matrix . Vector v2 = org . la4j . matrix . MatrixTest . v ( 0.0 , 1.0 , 0.0 ) ; org . la4j . matrix . Vector v3 = org . la4j . matrix . MatrixTest . v ( 0.0 , 0.0 , 1.0 ) ; org . la4j . matrix . Matrix b = mz ( 0 , 0 ) ; b = b . insertRow ( 0 , v3 ) ; b = b . insertRow ( 0 , v2 ) ; b = b . insertRow ( 0 , v1 ) ; "<AssertPlaceHolder>" ; } mz ( int , int ) { return org . la4j . Matrix . zero ( rows , columns ) ; }
org . junit . Assert . assertEquals ( a , b )
getAsString_null ( ) { javax . faces . component . UIComponent component = org . oscm . ui . converter . ConverterTestHelper . getComponent ( false , null , null , "discount" ) ; java . lang . String actual = converter . getAsString ( context , component , null ) ; "<AssertPlaceHolder>" ; } getAsString ( javax . faces . context . FacesContext , javax . faces . component . UIComponent , java . lang . Object ) { java . lang . String retVal ; if ( ! ( object instanceof org . oscm . internal . vo . VOPaymentInfo ) ) { retVal = "" ; } else { retVal = java . lang . String . valueOf ( ( ( org . oscm . internal . vo . VOPaymentInfo ) ( object ) ) . getKey ( ) ) ; } return retVal ; }
org . junit . Assert . assertNull ( actual )
testGetBusinessObjectDataUploadCredential ( ) { org . finra . herd . model . api . xml . BusinessObjectDataKey businessObjectDataKey = new org . finra . herd . model . api . xml . BusinessObjectDataKey ( BDEF_NAMESPACE , BDEF_NAME , FORMAT_USAGE_CODE , FORMAT_FILE_TYPE_CODE , FORMAT_VERSION , PARTITION_VALUE , SUBPARTITION_VALUES , DATA_VERSION ) ; java . lang . String delimitedSubPartitionValues = java . lang . String . join ( "|" , org . finra . herd . rest . SUBPARTITION_VALUES ) ; org . finra . herd . model . api . xml . AwsCredential awsCredential = new org . finra . herd . model . api . xml . AwsCredential ( AWS_ASSUMED_ROLE_ACCESS_KEY , AWS_ASSUMED_ROLE_SECRET_KEY , AWS_ASSUMED_ROLE_SESSION_TOKEN , AWS_ASSUMED_ROLE_SESSION_EXPIRATION_TIME ) ; org . finra . herd . model . api . xml . StorageUnitUploadCredential storageUnitUploadCredential = new org . finra . herd . model . api . xml . StorageUnitUploadCredential ( awsCredential , AWS_KMS_KEY_ID ) ; when ( herdStringHelper . splitStringWithDefaultDelimiterEscaped ( delimitedSubPartitionValues ) ) . thenReturn ( org . finra . herd . rest . SUBPARTITION_VALUES ) ; when ( storageUnitService . getStorageUnitUploadCredential ( businessObjectDataKey , org . finra . herd . rest . CREATE_NEW_VERSION , org . finra . herd . rest . STORAGE_NAME ) ) . thenReturn ( storageUnitUploadCredential ) ; org . finra . herd . model . api . xml . BusinessObjectDataUploadCredential result = businessObjectDataRestController . getBusinessObjectDataUploadCredential ( org . finra . herd . rest . BDEF_NAMESPACE , org . finra . herd . rest . BDEF_NAME , org . finra . herd . rest . FORMAT_USAGE_CODE , org . finra . herd . rest . FORMAT_FILE_TYPE_CODE , org . finra . herd . rest . FORMAT_VERSION , org . finra . herd . rest . PARTITION_VALUE , org . finra . herd . rest . DATA_VERSION , org . finra . herd . rest . CREATE_NEW_VERSION , org . finra . herd . rest . STORAGE_NAME , delimitedSubPartitionValues ) ; verify ( herdStringHelper ) . splitStringWithDefaultDelimiterEscaped ( delimitedSubPartitionValues ) ; verify ( storageUnitService ) . getStorageUnitUploadCredential ( businessObjectDataKey , org . finra . herd . rest . CREATE_NEW_VERSION , org . finra . herd . rest . STORAGE_NAME ) ; verifyNoMoreInteractionsHelper ( ) ; "<AssertPlaceHolder>" ; } verifyNoMoreInteractionsHelper ( ) { verifyNoMoreInteractions ( awsHelper , javaPropertiesHelper , retryPolicyFactory , s3Operations ) ; }
org . junit . Assert . assertEquals ( new org . finra . herd . model . api . xml . BusinessObjectDataUploadCredential ( awsCredential , AWS_KMS_KEY_ID ) , result )
should_require_0_to_100_multiplers_to_be_set ( ) { java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; for ( int i = 0 ; i <= 100 ; i ++ ) { sb . append ( i ) ; sb . append ( '\t' ) ; sb . append ( ( i + 0.5 ) ) ; sb . append ( '\n' ) ; } com . google . common . io . Files . write ( B ( sb . toString ( ) ) , input ) ; au . edu . wehi . idsv . PrecomputedGcBiasAdjuster gcba = new au . edu . wehi . idsv . PrecomputedGcBiasAdjuster ( input ) ; for ( int i = 0 ; i <= 100 ; i ++ ) { "<AssertPlaceHolder>" ; } } adjustmentMultiplier ( int ) { if ( gcPercentage < 0 ) throw new java . lang . IllegalArgumentException ( "GC<sp>percentage<sp>cannot<sp>be<sp>negative" ) ; if ( gcPercentage > 100 ) throw new java . lang . IllegalArgumentException ( "GC<sp>percentage<sp>cannot<sp>be<sp>greater<sp>than<sp>100" ) ; return adjustment [ gcPercentage ] ; }
org . junit . Assert . assertEquals ( ( i + 0.5 ) , gcba . adjustmentMultiplier ( i ) , 0 )
testForUnsignedTimeArray ( ) { java . sql . Time [ ] timeArr = new java . sql . Time [ 2 ] ; timeArr [ 0 ] = new java . sql . Time ( java . lang . System . currentTimeMillis ( ) ) ; timeArr [ 1 ] = new java . sql . Time ( 900000L ) ; org . apache . phoenix . schema . types . PhoenixArray arr = org . apache . phoenix . schema . types . PArrayDataType . instantiatePhoenixArray ( PUnsignedTime . INSTANCE , timeArr ) ; PUnsignedTimeArray . INSTANCE . toObject ( arr , PUnsignedTimeArray . INSTANCE ) ; byte [ ] bytes = PUnsignedTimeArray . INSTANCE . toBytes ( arr ) ; org . apache . phoenix . schema . types . PhoenixArray resultArr = ( ( org . apache . phoenix . schema . types . PhoenixArray ) ( PUnsignedTimeArray . INSTANCE . toObject ( bytes , 0 , bytes . length ) ) ) ; "<AssertPlaceHolder>" ; } toObject ( java . lang . Object , org . apache . phoenix . schema . types . PDataType , org . apache . phoenix . schema . SortOrder ) { return toObject ( object , actualType ) ; }
org . junit . Assert . assertEquals ( arr , resultArr )
testSingleResultFromDeviceIndex ( ) { "<AssertPlaceHolder>" ; { org . hibernate . search . FullTextSession fullTextSession = node . openFullTextSession ( ) ; org . hibernate . Transaction transaction = fullTextSession . beginTransaction ( ) ; org . infinispan . hibernate . search . sharedIndex . Toaster toaster = new org . infinispan . hibernate . search . sharedIndex . Toaster ( "A1" ) ; fullTextSession . save ( toaster ) ; transaction . commit ( ) ; fullTextSession . close ( ) ; verifyResult ( node ) ; } } clusterSize ( org . hibernate . search . test . util . FullTextSessionBuilder , java . lang . Class ) { org . hibernate . search . spi . SearchIntegrator integrator = node . getSearchFactory ( ) . unwrap ( org . hibernate . search . spi . SearchIntegrator . class ) ; org . hibernate . search . engine . spi . EntityIndexBinding indexBinding = integrator . getIndexBinding ( org . infinispan . hibernate . search . sharedIndex . SharedIndexTest . TOASTER_TYPE ) ; org . hibernate . search . indexes . spi . DirectoryBasedIndexManager indexManager = ( ( org . hibernate . search . indexes . spi . DirectoryBasedIndexManager ) ( indexBinding . getIndexManagerSelector ( ) . all ( ) . iterator ( ) . next ( ) ) ) ; org . infinispan . hibernate . search . spi . InfinispanDirectoryProvider directoryProvider = ( ( org . infinispan . hibernate . search . spi . InfinispanDirectoryProvider ) ( indexManager . getDirectoryProvider ( ) ) ) ; org . infinispan . manager . EmbeddedCacheManager cacheManager = directoryProvider . getCacheManager ( ) ; java . util . List < org . infinispan . remoting . transport . Address > members = cacheManager . getMembers ( ) ; return members . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , clusterSize ( node , org . infinispan . hibernate . search . sharedIndex . Toaster . class ) )
testSetOrderableChildren ( ) { org . apache . jackrabbit . oak . api . Tree tree = root . getTree ( "/" ) . addChild ( "test2" ) ; tree . setOrderableChildren ( true ) ; java . lang . String [ ] childNames = new java . lang . String [ ] { "a" , "b" , "c" , "d" } ; for ( java . lang . String name : childNames ) { tree . addChild ( name ) ; } int index = 0 ; for ( org . apache . jackrabbit . oak . api . Tree child : tree . getChildren ( ) ) { "<AssertPlaceHolder>" ; } } getName ( ) { return name ; }
org . junit . Assert . assertEquals ( childNames [ ( index ++ ) ] , child . getName ( ) )
testRollupSetting ( ) { java . util . List < org . joda . time . Interval > intervals = com . google . common . collect . Lists . newArrayList ( org . apache . druid . java . util . common . Intervals . of ( "2012-01-08T00Z/2012-01-11T00Z" ) , org . apache . druid . java . util . common . Intervals . of ( "2012-01-07T00Z/2012-01-08T00Z" ) , org . apache . druid . java . util . common . Intervals . of ( "2012-01-03T00Z/2012-01-04T00Z" ) , org . apache . druid . java . util . common . Intervals . of ( "2012-01-01T00Z/2012-01-03T00Z" ) ) ; final org . apache . druid . segment . indexing . granularity . GranularitySpec spec = new org . apache . druid . segment . indexing . granularity . UniformGranularitySpec ( org . apache . druid . java . util . common . granularity . Granularities . DAY , org . apache . druid . java . util . common . granularity . Granularities . NONE , false , intervals ) ; "<AssertPlaceHolder>" ; } isRollup ( ) { return rollup ; }
org . junit . Assert . assertFalse ( spec . isRollup ( ) )
failureThenSecondAlreadySucceeded ( ) { initReg . remove ( ) ; initPair ( first , jetbrains . jetpad . base . Asyncs . constant ( "a" ) ) ; first . failure ( new java . lang . Throwable ( ) ) ; "<AssertPlaceHolder>" ; } failure ( java . lang . Throwable ) { if ( alreadyHandled ( ) ) { throw new java . lang . IllegalStateException ( "Async<sp>already<sp>completed" ) ; } myFailureThrowable = throwable ; myFailed = true ; for ( jetbrains . jetpad . base . function . Consumer < java . lang . Throwable > handler : myFailureHandlers ) { try { handler . accept ( throwable ) ; } catch ( java . lang . Throwable t ) { jetbrains . jetpad . base . ThrowableHandlers . handle ( t ) ; } } clearHandlers ( ) ; }
org . junit . Assert . assertNotNull ( error )
testDefaultMaxSize ( ) { org . pm4j . common . util . collection . LRUMap < java . lang . Integer , java . lang . String > lruMap = new org . pm4j . common . util . collection . LRUMap < java . lang . Integer , java . lang . String > ( ) ; for ( java . lang . Integer i = 0 ; i < 1001 ; i ++ ) { lruMap . put ( i , java . lang . String . valueOf ( i ) ) ; } "<AssertPlaceHolder>" ; } size ( ) { return internalMap . size ( ) ; }
org . junit . Assert . assertEquals ( 1000 , lruMap . size ( ) )
testCopyConfigurationAliased ( ) { java . lang . Boolean configurationAliasRegistered = false ; java . lang . Class configurationClass = it . geosolutions . geobatch . actions . commons . CopyConfiguration . class ; java . util . Iterator < java . util . Map . Entry < java . lang . String , java . lang . Class < ? > > > it = aliasRegistry . iterator ( ) ; while ( it . hasNext ( ) ) { java . util . Map . Entry < java . lang . String , java . lang . Class < ? > > alias = it . next ( ) ; it . geosolutions . geobatch . annotations . Action action = ( ( it . geosolutions . geobatch . annotations . Action ) ( it . geosolutions . geobatch . actions . commons . CopyAction . class . getAnnotations ( ) [ 0 ] ) ) ; java . lang . String configurationAlias = action . configurationAlias ( ) ; if ( ( configurationAlias == null ) || ( configurationAlias . isEmpty ( ) ) ) { configurationAlias = configurationClass . getSimpleName ( ) ; } if ( ( alias . getKey ( ) . equals ( configurationAlias ) ) && ( alias . getValue ( ) . equals ( configurationClass ) ) ) { configurationAliasRegistered = true ; break ; } } "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) { return true ; } if ( ! ( super . equals ( obj ) ) ) { return false ; } if ( ( it . geosolutions . geobatch . flow . event . generator . file . FileBasedEventGenerator . getClass ( ) ) != ( obj . getClass ( ) ) ) { return false ; } it . geosolutions . geobatch . flow . event . generator . file . FileBasedEventGenerator < it . geosolutions . filesystemmonitor . monitor . FileSystemEvent > other = null ; if ( obj instanceof it . geosolutions . geobatch . flow . event . generator . file . FileBasedEventGenerator ) { other = ( ( it . geosolutions . geobatch . flow . event . generator . file . FileBasedEventGenerator ) ( obj ) ) ; } else { throw new java . lang . IllegalArgumentException ( "The<sp>object<sp>is<sp>not<sp>a<sp>FileBasedEventGenerator." ) ; } if ( ( fsMonitor ) == null ) { if ( ( other . fsMonitor ) != null ) { return false ; } } else if ( ! ( fsMonitor . equals ( other . fsMonitor ) ) ) { return false ; } if ( ( watchDirectory ) == null ) { if ( ( other . watchDirectory ) != null ) { return false ; } } else if ( ! ( watchDirectory . equals ( other . watchDirectory ) ) ) { return false ; } if ( ( wildCard ) == null ) { if ( ( other . wildCard ) != null ) { return false ; } } else if ( ! ( wildCard . equals ( other . wildCard ) ) ) { return false ; } return true ; }
org . junit . Assert . assertTrue ( configurationAliasRegistered )
testFetchByPrimaryKeysWithNoPrimaryKeys ( ) { java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; java . util . Map < java . io . Serializable , com . liferay . portal . workflow . kaleo . model . KaleoTransition > kaleoTransitions = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( kaleoTransitions . isEmpty ( ) )
noInterceptors ( ) { "<AssertPlaceHolder>" ; } getInterceptorFactories ( ) { return copyOf ( interceptorFactories ) ; }
org . junit . Assert . assertThat ( manager . getInterceptorFactories ( ) , empty ( ) )
registerConverterClass ( ) { com . basho . riak . client . api . convert . ConverterFactory factory = com . basho . riak . client . api . convert . ConverterFactory . getInstance ( ) ; factory . registerConverterForClass ( com . basho . riak . client . api . convert . ConverterFactoryTest . Pojo . class , new com . basho . riak . client . api . convert . ConverterFactoryTest . MyConverter ( ) ) ; com . basho . riak . client . api . convert . Converter < com . basho . riak . client . api . convert . ConverterFactoryTest . Pojo > converter = factory . getConverter ( com . basho . riak . client . api . convert . ConverterFactoryTest . Pojo . class ) ; "<AssertPlaceHolder>" ; } getConverter ( com . fasterxml . jackson . core . type . TypeReference ) { return getConverter ( null , typeReference ) ; }
org . junit . Assert . assertTrue ( ( converter instanceof com . basho . riak . client . api . convert . ConverterFactoryTest . MyConverter ) )
shouldNotFailOnEmptyQueue ( ) { java . nio . file . Path path = net . openhft . chronicle . queue . reader . DirectoryUtils . tempDir ( "shouldNotFailOnEmptyQueue" ) . toPath ( ) ; path . toFile ( ) . mkdirs ( ) ; new net . openhft . chronicle . queue . reader . ChronicleReader ( ) . withBasePath ( path ) . withMessageSink ( capturedOutput :: add ) . execute ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return stores . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( capturedOutput . isEmpty ( ) )
testRemoveUmlPrefixes2 ( ) { java . lang . String fullQualifiedName = "Data::de.crowdcode.test.Company" ; java . lang . String result = packageHelper . removeUmlPrefixes ( fullQualifiedName ) ; "<AssertPlaceHolder>" ; } removeUmlPrefixes ( java . lang . String ) { java . lang . String result = fullQualifiedName . replace ( UmlTypePrefix . MAGIC_DRAW_PROFILE_DATATYPES . getValue ( ) , "" ) ; result = result . replace ( UmlTypePrefix . UML_PRIMITIVE_TYPES . getValue ( ) , "" ) ; result = result . replace ( UmlTypePrefix . JAVA_PRIMITIVE_TYPES . getValue ( ) , "" ) ; result = result . replace ( UmlTypePrefix . VALIDATION_PROFILE_OCL_LIBRARY . getValue ( ) , "" ) ; if ( org . apache . commons . lang3 . StringUtils . contains ( result , UmlTypePrefix . DATA_DATATYPE . getValue ( ) ) ) { result = result . replace ( UmlTypePrefix . DATA_DATATYPE . getValue ( ) , "" ) ; } if ( org . apache . commons . lang3 . StringUtils . contains ( result , UmlTypePrefix . DATA_DATATYPE_BINDINGS . getValue ( ) ) ) { result = result . replace ( UmlTypePrefix . DATA_DATATYPE_BINDINGS . getValue ( ) , "" ) ; } return result ; }
org . junit . Assert . assertEquals ( "Data::de.crowdcode.test.Company" , result )
shouldReturnTrueWhenEntityWithGivenProductAndCompanyDoesnotExistsInDB ( ) { com . qcadoo . model . api . search . SearchCriterion criterion1 = com . qcadoo . model . api . search . SearchRestrictions . eq ( com . qcadoo . mes . productCatalogNumbers . hooks . CATALOG_NUMBER , entity . getStringField ( com . qcadoo . mes . productCatalogNumbers . hooks . CATALOG_NUMBER ) ) ; com . qcadoo . model . api . search . SearchCriterion criterion2 = com . qcadoo . model . api . search . SearchRestrictions . belongsTo ( com . qcadoo . mes . productCatalogNumbers . hooks . PRODUCT , entity . getBelongsToField ( com . qcadoo . mes . productCatalogNumbers . hooks . PRODUCT ) ) ; given ( entity . getId ( ) ) . willReturn ( null ) ; given ( criteria . add ( criterion1 ) ) . willReturn ( criteria ) ; given ( criteria . add ( criterion2 ) ) . willReturn ( criteria ) ; given ( criteria . list ( ) ) . willReturn ( searchResult ) ; given ( searchResult . getEntities ( ) ) . willReturn ( productCatalogNumbers ) ; given ( productCatalogNumbers . isEmpty ( ) ) . willReturn ( true ) ; boolean result = productCatalogNumbersHooks . checkIfExistsCatalogNumberWithProductAndCompany ( dataDefinition , entity ) ; "<AssertPlaceHolder>" ; } checkIfExistsCatalogNumberWithProductAndCompany ( com . qcadoo . model . api . DataDefinition , com . qcadoo . model . api . Entity ) { com . qcadoo . model . api . search . SearchCriteriaBuilder criteria = dataDefinitionService . get ( ProductCatalogNumbersConstants . PLUGIN_IDENTIFIER , ProductCatalogNumbersConstants . MODEL_PRODUCT_CATALOG_NUMBERS ) . find ( ) . add ( com . qcadoo . model . api . search . SearchRestrictions . belongsTo ( com . qcadoo . mes . productCatalogNumbers . hooks . PRODUCT , entity . getBelongsToField ( com . qcadoo . mes . productCatalogNumbers . hooks . PRODUCT ) ) ) . add ( com . qcadoo . model . api . search . SearchRestrictions . belongsTo ( com . qcadoo . mes . productCatalogNumbers . hooks . COMPANY , entity . getBelongsToField ( com . qcadoo . mes . productCatalogNumbers . hooks . COMPANY ) ) ) ; if ( ( entity . getId ( ) ) != null ) { criteria . add ( com . qcadoo . model . api . search . SearchRestrictions . ne ( "id" , entity . getId ( ) ) ) ; } java . util . List < com . qcadoo . model . api . Entity > catalogsNumbers = criteria . list ( ) . getEntities ( ) ; if ( catalogsNumbers . isEmpty ( ) ) { return true ; } else { entity . addGlobalError ( "productCatalogNumbers.productCatalogNumber.validationError.alreadyExistsProductForCompany" ) ; return false ; } }
org . junit . Assert . assertTrue ( result )
testCreateCronJobTrigger ( ) { "<AssertPlaceHolder>" ; } createCronJobTrigger ( ) { return new org . pentaho . platform . plugin . services . importexport . exportManifest . bindings . CronJobTrigger ( ) ; }
org . junit . Assert . assertNotNull ( factory . createCronJobTrigger ( ) )
testOffsetRightAfterCommentSymbol ( ) { net . roboconf . tooling . core . autocompletion . Couple couple = prepare ( "app1" , "initial.instances" , 2 ) ; java . util . List < net . roboconf . tooling . core . autocompletion . ICompletionProposer . RoboconfCompletionProposal > proposals = couple . proposer . findProposals ( couple . text ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . map . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , proposals . size ( ) )
testReportForFileSizeGroupByPUIDWithFilter ( ) { reportData = reportDao . getReportData ( filter , ReportFieldEnum . FILE_SIZE , getGroupByFieldList ( ReportFieldEnum . PUID ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return size ; }
org . junit . Assert . assertTrue ( ( ( reportData . size ( ) ) > 0 ) )
shouldProcessTestClasses ( ) { java . lang . System . setProperty ( "test" , "*Test1" ) ; final java . io . File testClassFile = new java . io . File ( ( ( ( com . github . bmsantos . core . cola . main . ColaMainTest . TARGET_DIR ) + ( separator ) ) + ( testClass1 ) ) ) ; final long initialSize = testClassFile . length ( ) ; uut . execute ( provider ) ; final long finalSize = testClassFile . length ( ) ; "<AssertPlaceHolder>" ; } execute ( com . github . bmsantos . core . cola . provider . IColaProvider ) { this . provider = provider ; failures = new java . util . ArrayList ( ) ; if ( ! ( isSet ( provider ) ) ) { return ; } final java . util . List < java . lang . String > targetClasses = provider . getTargetClasses ( ) ; if ( ! ( isSet ( targetClasses ) ) ) { return ; } for ( final java . lang . String className : targetClasses ) { try { processClass ( className , null ) ; } catch ( final java . lang . Throwable t ) { com . github . bmsantos . core . cola . main . ColaMain . log . error ( java . lang . String . format ( config . error ( "failed.process.file" ) , className ) , t ) ; failures . add ( java . lang . String . format ( config . error ( "failed.processing" ) , className , t . getMessage ( ) ) ) ; } } if ( ! ( failures . isEmpty ( ) ) ) { com . github . bmsantos . core . cola . main . ColaMain . log . error ( java . lang . String . format ( config . error ( "failed.tests" ) , failures . size ( ) , targetClasses . size ( ) ) ) ; for ( final java . lang . String failure : failures ) { com . github . bmsantos . core . cola . main . ColaMain . log . error ( failure ) ; } throw new com . github . bmsantos . core . cola . exceptions . ColaExecutionException ( config . error ( "processing" ) ) ; } }
org . junit . Assert . assertThat ( ( initialSize < finalSize ) , org . hamcrest . Matchers . is ( true ) )
testDocFullUrlCreation ( ) { org . nuxeo . ecm . platform . util . RepositoryLocation rep = new org . nuxeo . ecm . platform . util . RepositoryLocation ( "alpha" ) ; org . nuxeo . ecm . core . api . DocumentRef docRef = new org . nuxeo . ecm . core . api . IdRef ( "xxxx-xxx-xxxxx-xxxxxxxxx-xx" ) ; final java . lang . String result = "http://localhost:8080/nuxeo/" + ( org . nuxeo . ecm . platform . ui . web . util . DocumentLocator . getFullDocumentUrl ( rep , docRef ) ) ; org . nuxeo . ecm . platform . ui . web . util . TestDocumentLocator . log . info ( ( "result:<sp>" + result ) ) ; try { "<AssertPlaceHolder>" ; } catch ( java . net . MalformedURLException e ) { org . junit . Assert . fail ( ( "not<sp>a<sp>valid<sp>result:<sp>" + ( e . getMessage ( ) ) ) ) ; } } info ( java . lang . String ) { facesMessages . add ( org . nuxeo . ecm . platform . signature . web . sign . INFO , getMessage ( msg ) ) ; }
org . junit . Assert . assertNotNull ( new java . net . URL ( result ) )
testGetCellScanner ( ) { java . util . List < org . apache . hadoop . hbase . Cell > a = new java . util . ArrayList ( ) ; org . apache . hadoop . hbase . KeyValue akv = new org . apache . hadoop . hbase . KeyValue ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "a" ) , ( - 1L ) ) ; a . add ( akv ) ; a . add ( new org . apache . hadoop . hbase . KeyValue ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "aa" ) , ( - 1L ) ) ) ; a . add ( new org . apache . hadoop . hbase . KeyValue ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "aaa" ) , ( - 1L ) ) ) ; java . util . List < org . apache . hadoop . hbase . Cell > b = new java . util . ArrayList ( ) ; org . apache . hadoop . hbase . KeyValue bkv = new org . apache . hadoop . hbase . KeyValue ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "b" ) , ( - 1L ) ) ; a . add ( bkv ) ; java . util . List < org . apache . hadoop . hbase . Cell > c = new java . util . ArrayList ( ) ; org . apache . hadoop . hbase . KeyValue ckv = new org . apache . hadoop . hbase . KeyValue ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "c" ) , ( - 1L ) ) ; c . add ( ckv ) ; java . util . List < java . util . List < ? extends org . apache . hadoop . hbase . Cell > > all = new java . util . ArrayList ( ) ; all . add ( a ) ; all . add ( b ) ; all . add ( c ) ; org . apache . hadoop . hbase . CellScanner scanner = org . apache . hadoop . hbase . protobuf . ReplicationProtbufUtil . getCellScanner ( all , 0 ) ; testAdvancetHasSameRow ( scanner , akv ) ; scanner . advance ( ) ; scanner . advance ( ) ; testAdvancetHasSameRow ( scanner , bkv ) ; testAdvancetHasSameRow ( scanner , ckv ) ; "<AssertPlaceHolder>" ; } advance ( ) { if ( ( this . currentIterator ) == null ) { if ( ! ( this . entries . hasNext ( ) ) ) return false ; this . currentIterator = this . entries . next ( ) . iterator ( ) ; } if ( this . currentIterator . hasNext ( ) ) { this . currentCell = this . currentIterator . next ( ) ; return true ; } this . currentCell = null ; this . currentIterator = null ; return advance ( ) ; }
org . junit . Assert . assertFalse ( scanner . advance ( ) )
testRooms ( ) { java . util . List < com . sap . core . odata . ref . model . Room > list = java . util . Arrays . asList ( new com . sap . core . odata . ref . model . Room ( 1 , null ) , new com . sap . core . odata . ref . model . Room ( 2 , null ) , new com . sap . core . odata . ref . model . Room ( 3 , null ) ) ; com . sap . core . odata . ref . model . Building building1 = new com . sap . core . odata . ref . model . Building ( 1 , null ) ; building1 . getRooms ( ) . add ( list . get ( 0 ) ) ; building1 . getRooms ( ) . add ( list . get ( 1 ) ) ; building1 . getRooms ( ) . add ( list . get ( 2 ) ) ; "<AssertPlaceHolder>" ; } getRooms ( ) { return rooms ; }
org . junit . Assert . assertEquals ( list , building1 . getRooms ( ) )
peekEmptyTest ( ) { final com . conversantmedia . util . concurrent . ConcurrentStack < java . lang . Integer > stack = new com . conversantmedia . util . concurrent . ConcurrentStack ( 10 ) ; "<AssertPlaceHolder>" ; } peek ( ) { int spin = 0 ; for ( ; ; ) { final long readLock = seqLock . readLock ( ) ; final int stackTop = this . stackTop . get ( ) ; if ( stackTop > 0 ) { final N n = stack . get ( ( stackTop - 1 ) ) ; if ( seqLock . readLockHeld ( readLock ) ) { return n ; } } else { return null ; } spin = com . conversantmedia . util . concurrent . Condition . progressiveYield ( spin ) ; } }
org . junit . Assert . assertNull ( stack . peek ( ) )
testgetDimensions ( ) { final org . apache . druid . segment . incremental . IncrementalIndex < org . apache . druid . query . aggregation . Aggregator > incrementalIndex = new org . apache . druid . segment . incremental . IncrementalIndex . Builder ( ) . setIndexSchema ( new org . apache . druid . segment . incremental . IncrementalIndexSchema . Builder ( ) . withMetrics ( new org . apache . druid . query . aggregation . CountAggregatorFactory ( "count" ) ) . withDimensionsSpec ( new org . apache . druid . data . input . impl . DimensionsSpec ( org . apache . druid . data . input . impl . DimensionsSpec . getDefaultSchemas ( java . util . Arrays . asList ( "dim0" , "dim1" ) ) , null , null ) ) . build ( ) ) . setMaxRowCount ( 1000000 ) . buildOnheap ( ) ; closerRule . closeLater ( incrementalIndex ) ; "<AssertPlaceHolder>" ; } getDimensionNames ( ) { return baseAdapter . getDimensionNames ( ) ; }
org . junit . Assert . assertEquals ( java . util . Arrays . asList ( "dim0" , "dim1" ) , incrementalIndex . getDimensionNames ( ) )
filestatFailed ( ) { jnr . posix . FileStat stat = null ; try { stat = jnr . posix . FileStatTest . posix . stat ( jnr . posix . FileStatTest . NON_EXISTENT_FILENAME ) ; } catch ( java . lang . UnsupportedOperationException e ) { } "<AssertPlaceHolder>" ; } stat ( java . lang . String ) { jnr . posix . FileStat stat = allocateStat ( ) ; if ( ( stat ( path , stat ) ) < 0 ) handler . error ( jnr . constants . platform . Errno . valueOf ( errno ( ) ) , "stat" , path ) ; return stat ; }
org . junit . Assert . assertTrue ( ( stat == null ) )
testIteratorNextWithCount ( ) { nom . tam . util . HashedList < nom . tam . util . HashedListTest . TestCursorValue > h = new nom . tam . util . HashedList < nom . tam . util . HashedListTest . TestCursorValue > ( ) ; h . add ( new nom . tam . util . HashedListTest . TestCursorValue ( "K1" , "R1" ) ) ; h . add ( new nom . tam . util . HashedListTest . TestCursorValue ( "K2" , "R2" ) ) ; h . add ( new nom . tam . util . HashedListTest . TestCursorValue ( "K3" , "R3" ) ) ; nom . tam . util . HashedListTest . TestCursorValue cursorValue = h . iterator ( 0 ) . next ( 2 ) ; "<AssertPlaceHolder>" ; } next ( int ) { for ( int index = 1 ; index < count ; index ++ ) { next ( ) ; } return next ( ) ; }
org . junit . Assert . assertEquals ( "K2" , cursorValue . key )
testListUsers ( ) { java . util . List < com . aliyun . odps . security . User > list = sm . listUsers ( ) ; for ( com . aliyun . odps . security . User user : list ) { "<AssertPlaceHolder>" ; user . reload ( ) ; } } getID ( ) { return model . id ; }
org . junit . Assert . assertNotNull ( user . getID ( ) )
inverse_style_downbond_ambiguous ( ) { org . openscience . cdk . interfaces . IAtomContainer m = new org . openscience . cdk . silent . AtomContainer ( 6 , 4 , 0 , 0 ) ; m . addAtom ( org . openscience . cdk . stereo . StereoElementFactoryTest . atom ( "O" , 1 , ( - 0.46 ) , 1.98 ) ) ; m . addAtom ( org . openscience . cdk . stereo . StereoElementFactoryTest . atom ( "C" , 1 , ( - 1.28 ) , 1.96 ) ) ; m . addAtom ( org . openscience . cdk . stereo . StereoElementFactoryTest . atom ( "C" , 1 , ( - 1.71 ) , 2.67 ) ) ; m . addAtom ( org . openscience . cdk . stereo . StereoElementFactoryTest . atom ( "C" , 3 , ( - 1.68 ) , 1.24 ) ) ; m . addAtom ( org . openscience . cdk . stereo . StereoElementFactoryTest . atom ( "C" , 3 , ( - 2.53 ) , 2.66 ) ) ; m . addAtom ( org . openscience . cdk . stereo . StereoElementFactoryTest . atom ( "O" , 1 , ( - 1.31 ) , 3.39 ) ) ; m . addBond ( 1 , 0 , IBond . Order . SINGLE ) ; m . addBond ( 1 , 2 , IBond . Order . SINGLE , IBond . Stereo . DOWN_INVERTED ) ; m . addBond ( 1 , 3 , IBond . Order . SINGLE ) ; m . addBond ( 2 , 4 , IBond . Order . SINGLE ) ; m . addBond ( 2 , 5 , IBond . Order . SINGLE ) ; org . openscience . cdk . stereo . StereoElementFactory factory = org . openscience . cdk . stereo . StereoElementFactory . using2DCoordinates ( m ) ; org . openscience . cdk . interfaces . ITetrahedralChirality element = factory . createTetrahedral ( m . getAtom ( 1 ) , org . openscience . cdk . stereo . Stereocenters . of ( m ) ) ; "<AssertPlaceHolder>" ; } of ( org . openscience . cdk . interfaces . IAtomContainer ) { org . openscience . cdk . stereo . EdgeToBondMap bondMap = org . openscience . cdk . stereo . EdgeToBondMap . withSpaceFor ( container ) ; int [ ] [ ] g = org . openscience . cdk . graph . GraphUtil . toAdjList ( container , bondMap ) ; org . openscience . cdk . stereo . Stereocenters stereocenters = new org . openscience . cdk . stereo . Stereocenters ( container , g , bondMap ) ; stereocenters . checkSymmetry ( ) ; return stereocenters ; }
org . junit . Assert . assertNull ( element )
deveGerarXMLDeAcordoComOPadraoEstabelecidoBCAliquotaPercentual ( ) { final com . fincatto . documentofiscal . nfe400 . classes . nota . NFNotaInfoItemImpostoPISST impostoPISST = new com . fincatto . documentofiscal . nfe400 . classes . nota . NFNotaInfoItemImpostoPISST ( ) ; impostoPISST . setPercentualAliquota ( new java . math . BigDecimal ( "99.99" ) ) ; impostoPISST . setValorBaseCalculo ( new java . math . BigDecimal ( "999999999999.99" ) ) ; impostoPISST . setValorTributo ( new java . math . BigDecimal ( "999999999999.99" ) ) ; final java . lang . String xmlEsperado = "<NFNotaInfoItemImpostoPISST><vBC>999999999999.99</vBC><pPIS>99.99</pPIS><vPIS>999999999999.99</vPIS></NFNotaInfoItemImpostoPISST>" ; "<AssertPlaceHolder>" ; } toString ( ) { return this . getDescricao ( ) ; }
org . junit . Assert . assertEquals ( xmlEsperado , impostoPISST . toString ( ) )
two_$parent_project_field_without_recursion ( ) { com . redhat . lightblue . query . Projection p = com . redhat . lightblue . eval . EvalTestContext . projectionFromJson ( "[{'field':'field6.nf7.$parent.$parent.field2'},{'field':'field6.nf7.$parent.$parent.field6.*'}]" ) ; com . redhat . lightblue . eval . Projector projector = com . redhat . lightblue . eval . Projector . getInstance ( p , md ) ; com . fasterxml . jackson . databind . JsonNode expectedNode = com . redhat . lightblue . util . JsonUtils . json ( "{'field2':'value2','field6':{'nf1':'nvalue1','nf2':'nvalue2','nf3':4,'nf4':false,'nf11':null}}" . replace ( '\'' , '\"' ) ) ; com . redhat . lightblue . util . JsonDoc pdoc = projector . project ( jsonDoc , com . redhat . lightblue . eval . JSON_NODE_FACTORY ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ops [ 0 ] ; }
org . junit . Assert . assertEquals ( expectedNode . toString ( ) , pdoc . toString ( ) )
testStdev0 ( ) { double [ ] [ ] ind = new double [ ] [ ] { new double [ ] { 5.1 , 3.5 , 1.4 } , new double [ ] { 4.9 , 3.0 , 1.4 } , new double [ ] { 4.7 , 3.2 , 1.3 } } ; org . nd4j . linalg . api . ndarray . INDArray in = org . nd4j . linalg . factory . Nd4j . create ( ind ) ; org . nd4j . linalg . api . ndarray . INDArray stdev = in . std ( 0 ) ; org . nd4j . linalg . api . ndarray . INDArray exp = org . nd4j . linalg . factory . Nd4j . create ( new double [ ] { 0.19999999999999973 , 0.2516611478423583 , 0.057735026918962505 } ) ; "<AssertPlaceHolder>" ; } create ( float [ ] , int [ ] , long ) { shape = org . nd4j . linalg . factory . Nd4j . getEnsuredShape ( shape ) ; if ( ( shape . length ) == 1 ) { if ( ( shape [ 0 ] ) == ( data . length ) ) { shape = new int [ ] { 1 , data . length } ; } else throw new org . nd4j . linalg . exception . ND4JIllegalStateException ( ( ( ( "Shape<sp>of<sp>the<sp>new<sp>array<sp>" + ( org . nd4j . linalg . factory . Arrays . toString ( shape ) ) ) + "<sp>doesn't<sp>match<sp>data<sp>length:<sp>" ) + ( data . length ) ) ) ; } org . nd4j . linalg . factory . Nd4j . checkShapeValues ( data . length , shape ) ; org . nd4j . linalg . factory . INDArray ret = org . nd4j . linalg . factory . Nd4j . INSTANCE . create ( data , shape , offset , org . nd4j . linalg . factory . Nd4j . order ( ) ) ; org . nd4j . linalg . factory . Nd4j . logCreationIfNecessary ( ret ) ; return ret ; }
org . junit . Assert . assertEquals ( exp , stdev )
testAllConstructorsStringGeneration ( ) { org . kie . workbench . common . services . datamodeller . core . DataModel dataModel = dataModelOracleDriver . createModel ( ) ; org . kie . workbench . common . services . datamodeller . core . DataObject object1 = dataModel . addDataObject ( "com.test.Object1" ) ; org . kie . workbench . common . services . datamodeller . core . DataObject object2 = dataModel . addDataObject ( "com.test.sub.Object2" ) ; org . kie . workbench . common . services . datamodeller . core . ObjectProperty prop1 = object1 . addProperty ( "attribute1" , "java.lang.String" ) ; ( ( org . kie . workbench . common . services . datamodeller . core . impl . ObjectPropertyImpl ) ( prop1 ) ) . setFileOrder ( 0 ) ; org . kie . workbench . common . services . datamodeller . core . ObjectProperty prop2 = object1 . addProperty ( "attribute2" , "long" 2 ) ; ( ( org . kie . workbench . common . services . datamodeller . core . impl . ObjectPropertyImpl ) ( prop2 ) ) . setFileOrder ( 1 ) ; org . kie . workbench . common . services . datamodeller . core . ObjectProperty prop3 = object1 . addProperty ( "attribute3" , object2 . getClassName ( ) ) ; ( ( org . kie . workbench . common . services . datamodeller . core . impl . ObjectPropertyImpl ) ( prop3 ) ) . setFileOrder ( 2 ) ; org . kie . workbench . common . services . datamodeller . core . ObjectProperty prop4 = object1 . addProperty ( "long" 3 , "long" ) ; ( ( org . kie . workbench . common . services . datamodeller . core . impl . ObjectPropertyImpl ) ( prop4 ) ) . setFileOrder ( 3 ) ; org . kie . workbench . common . services . datamodeller . core . Annotation key = new org . kie . workbench . common . services . datamodeller . core . impl . AnnotationImpl ( annotationDefinitions . get ( org . kie . api . definition . type . Key . class . getName ( ) ) ) ; org . kie . workbench . common . services . datamodeller . core . Annotation position = new org . kie . workbench . common . services . datamodeller . core . impl . AnnotationImpl ( annotationDefinitions . get ( org . kie . api . definition . type . Position . class . getName ( ) ) ) ; position . setValue ( "value" , "0" ) ; prop3 . addAnnotation ( key ) ; prop3 . addAnnotation ( position ) ; position = new org . kie . workbench . common . services . datamodeller . core . impl . AnnotationImpl ( annotationDefinitions . get ( org . kie . api . definition . type . Position . class . getName ( ) ) ) ; position . setValue ( "value" , "long" 0 ) ; prop1 . addAnnotation ( key ) ; prop1 . addAnnotation ( position ) ; position = new org . kie . workbench . common . services . datamodeller . core . impl . AnnotationImpl ( annotationDefinitions . get ( org . kie . api . definition . type . Position . class . getName ( ) ) ) ; position . setValue ( "value" , "long" 1 ) ; prop2 . addAnnotation ( key ) ; prop2 . addAnnotation ( position ) ; position = new org . kie . workbench . common . services . datamodeller . core . impl . AnnotationImpl ( annotationDefinitions . get ( org . kie . api . definition . type . Position . class . getName ( ) ) ) ; position . setValue ( "value" , "3" ) ; prop4 . addAnnotation ( position ) ; org . kie . workbench . common . services . datamodeller . codegen . GenerationContext generationContext = new org . kie . workbench . common . services . datamodeller . codegen . GenerationContext ( dataModel ) ; try { java . lang . String result = engine . generateAllConstructorsString ( generationContext , object1 ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; } } getProperty ( java . lang . String ) { if ( name == null ) return null ; for ( org . kie . workbench . common . services . datamodeller . core . ObjectProperty property : properties ) { if ( name . equals ( property . getName ( ) ) ) { return property ; } } return null ; }
org . junit . Assert . assertEquals ( results . getProperty ( "long" 4 ) , result )
testBefore ( ) { final org . apache . drill . exec . expr . holders . VarCharHolder left = org . apache . drill . exec . fn . impl . TestByteComparisonFunctions . goodbye ; final org . apache . drill . exec . expr . holders . VarCharHolder right = org . apache . drill . exec . fn . impl . TestByteComparisonFunctions . hello ; "<AssertPlaceHolder>" ; } compare ( io . netty . buffer . DrillBuf , int , int , io . netty . buffer . DrillBuf , int , int ) { rangeCheck ( left , lStart , lEnd , right , rStart , rEnd ) ; return org . apache . drill . exec . expr . fn . impl . ByteFunctionHelpers . memcmp ( left . memoryAddress ( ) , lStart , lEnd , right . memoryAddress ( ) , rStart , rEnd ) ; }
org . junit . Assert . assertTrue ( ( ( org . apache . drill . exec . expr . fn . impl . ByteFunctionHelpers . compare ( left . buffer , left . start , left . end , right . buffer , right . start , right . end ) ) == ( - 1 ) ) )
testBadBatchSize ( ) { org . codice . ddf . commands . catalog . ConsoleOutput consoleOutput = new org . codice . ddf . commands . catalog . ConsoleOutput ( ) ; consoleOutput . interceptSystemOut ( ) ; org . codice . ddf . commands . catalog . RemoveAllCommand command = new org . codice . ddf . commands . catalog . RemoveAllCommand ( ) ; command . batchSize = 0 ; command . doExecute ( ) ; consoleOutput . resetSystemOut ( ) ; try { java . lang . String message = java . lang . String . format ( RemoveAllCommand . BATCH_SIZE_ERROR_MESSAGE_FORMAT , 0 ) ; java . lang . String expectedPrintOut = ( ( org . codice . ddf . commands . catalog . TestRemoveAllCommand . RED_CONSOLE_COLOR ) + message ) + ( org . codice . ddf . commands . catalog . TestRemoveAllCommand . DEFAULT_CONSOLE_COLOR ) ; "<AssertPlaceHolder>" ; } finally { consoleOutput . closeBuffer ( ) ; } } getOutput ( ) { return buffer . toString ( ) ; }
org . junit . Assert . assertThat ( consoleOutput . getOutput ( ) , org . hamcrest . Matchers . startsWith ( expectedPrintOut ) )
schemeTest ( ) { final java . lang . String expectedSchemeDescription = "CREATE<sp>TABLE<sp>IF<sp>NOT<sp>EXISTS<sp>\"RingOfPower\"<sp>(" + "id<sp>text,<sp>person<sp>list<text>,<sp>PRIMARY<sp>KEY<sp>((id)));" ; "<AssertPlaceHolder>" ; cassandraFactory . getSession ( ) . execute ( org . nohope . cassandra . mapservice . CCollectionListIT . SCHEME . buildTableDescription ( ) ) ; } buildTableDescription ( ) { return java . text . MessageFormat . format ( ( ( ( "CREATE<sp>TABLE<sp>IF<sp>NOT<sp>EXISTS<sp>\"{0}\"<sp>" + '(' ) + "{1},<sp>" ) + "PRIMARY<sp>KEY<sp>(({2}){3}));" ) , tableName , formColumns ( ) , org . nohope . cassandra . mapservice . TableScheme . fromColumns ( partitionKeys ) , ( clusteringKeys . isEmpty ( ) ? "" : ",<sp>" + ( org . nohope . cassandra . mapservice . TableScheme . fromColumns ( clusteringKeys ) ) ) ) ; }
org . junit . Assert . assertEquals ( expectedSchemeDescription , org . nohope . cassandra . mapservice . CCollectionListIT . SCHEME . buildTableDescription ( ) )
testCloneAndAugmentFlowWithSetVlanId ( ) { org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . flow . FlowConvertorTest . MockFlow mockFlow = new org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . flow . FlowConvertorTest . MockFlow ( ) ; org . opendaylight . yang . gen . v1 . urn . opendaylight . action . types . rev131112 . action . list . Action action1 = org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . flow . FlowConvertorTest . createAction ( new org . opendaylight . yang . gen . v1 . urn . opendaylight . action . types . rev131112 . action . action . SetVlanIdActionCaseBuilder ( ) . setSetVlanIdAction ( new org . opendaylight . yang . gen . v1 . urn . opendaylight . action . types . rev131112 . action . action . set . vlan . id . action . _case . SetVlanIdActionBuilder ( ) . setVlanId ( new org . opendaylight . yang . gen . v1 . urn . opendaylight . l2 . types . rev130827 . VlanId ( 10 ) ) . build ( ) ) . build ( ) , 0 ) ; mockFlow . setMatch ( new org . opendaylight . yang . gen . v1 . urn . opendaylight . flow . types . rev131026 . flow . MatchBuilder ( ) . setEthernetMatch ( org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . flow . FlowConvertorTest . createEthernetMatch ( ) ) . build ( ) ) ; mockFlow . setInstructions ( org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . flow . FlowConvertorTest . toApplyInstruction ( java . util . Collections . singletonList ( action1 ) ) ) ; org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . data . VersionDatapathIdConvertorData data = new org . opendaylight . openflowplugin . openflow . md . core . sal . convertor . data . VersionDatapathIdConvertorData ( org . opendaylight . openflowplugin . api . OFConstants . OFP_VERSION_1_3 ) ; data . setDatapathId ( BigInteger . ONE ) ; java . util . List < org . opendaylight . yang . gen . v1 . urn . opendaylight . openflow . protocol . rev130731 . FlowModInputBuilder > flowModInputBuilders = convert ( mockFlow , data ) ; "<AssertPlaceHolder>" ; } size ( ) { return flowRegistry . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , flowModInputBuilders . size ( ) )
hasDefinedAttribute ( ) { com . vaadin . flow . dom . Element e = com . vaadin . flow . dom . ElementFactory . createDiv ( ) ; e . setAttribute ( "foo" , "bar" ) ; "<AssertPlaceHolder>" ; } hasAttribute ( com . vaadin . flow . dom . Element ) { return ! ( element . getClassList ( ) . isEmpty ( ) ) ; }
org . junit . Assert . assertTrue ( e . hasAttribute ( "foo" ) )
testQueryDefaultValueInComplex ( ) { org . nuxeo . ecm . core . api . DocumentModel doc = session . createDocumentModel ( "/" , "doc" , "MyDocType" ) ; session . createDocument ( doc ) ; session . save ( ) ; org . nuxeo . ecm . core . api . DocumentModelList dml = session . query ( "SELECT<sp>*<sp>FROM<sp>MyDocType<sp>WHERE<sp>my:complex/testDefault<sp>=<sp>'the<sp>default<sp>value'" ) ; "<AssertPlaceHolder>" ; } size ( ) { return getCollectedDocumentIds ( ) . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , dml . size ( ) )
testCopyDynAny ( ) { java . lang . String msg ; org . omg . CORBA . TypeCode tc = null ; org . omg . DynamicAny . DynAny dynAny = null ; org . omg . DynamicAny . DynAny dynAny2 = null ; tc = orb . get_primitive_tc ( org . omg . CORBA . TCKind . tk_long ) ; dynAny = createDynAnyFromTypeCode ( tc ) ; try { dynAny . insert_long ( 700 ) ; } catch ( java . lang . Throwable ex ) { org . junit . Assert . fail ( ( "Failed<sp>to<sp>insert<sp>value<sp>into<sp>DynAny<sp>object:<sp>" + ex ) ) ; } dynAny2 = dynAny . copy ( ) ; msg = "The<sp>DynAny<sp>object<sp>created<sp>with<sp>the<sp>DynAny::copy<sp>operation<sp>" ; msg += "is<sp>not<sp>equal<sp>to<sp>the<sp>DynAny<sp>object<sp>it<sp>was<sp>copied<sp>from" ; "<AssertPlaceHolder>" ; } equal ( java . lang . Object ) { if ( ( obj1 == null ) || ( ( current ) == null ) ) { throw new org . jacorb . collection . util . ObjectInvalid ( ) ; } check_object ( obj1 ) ; return ops . equal ( current , ( ( org . omg . CORBA . Any ) ( obj1 ) ) ) ; }
org . junit . Assert . assertTrue ( msg , dynAny . equal ( dynAny2 ) )
testDiscoverUnmappedKeyspaces ( ) { cassandraClientHolder . getClient ( ) . system_add_keyspace ( setupOtherKeyspace ( configuration , "OtherKeyspace" , false ) ) ; java . util . List < org . apache . cassandra . thrift . KsDef > keyspaces = schemaManagerService . findUnmappedKeyspaces ( ) ; boolean foundCreated = false ; for ( org . apache . cassandra . thrift . KsDef ksDef : keyspaces ) { if ( org . apache . commons . lang . StringUtils . equals ( ksDef . name , "OtherKeyspace" ) ) { foundCreated = true ; break ; } } "<AssertPlaceHolder>" ; } findUnmappedKeyspaces ( ) { java . util . List < org . apache . cassandra . thrift . KsDef > defs ; try { defs = cassandraClientHolder . getClient ( ) . describe_keyspaces ( ) ; for ( java . util . Iterator < org . apache . cassandra . thrift . KsDef > iterator = defs . iterator ( ) ; iterator . hasNext ( ) ; ) { org . apache . cassandra . thrift . KsDef ksDef = iterator . next ( ) ; java . lang . String name = ksDef . name ; org . apache . cassandra . hadoop . hive . metastore . SchemaManagerService . log . debug ( "Found<sp>ksDef<sp>name:<sp>{}" , name ) ; if ( ( ( org . apache . commons . lang . StringUtils . indexOfAny ( name , org . apache . cassandra . hadoop . hive . metastore . SchemaManagerService . SYSTEM_KEYSPACES ) ) > ( - 1 ) ) || ( isKeyspaceMapped ( name ) ) ) { org . apache . cassandra . hadoop . hive . metastore . SchemaManagerService . log . debug ( "REMOVING<sp>ksDef<sp>name<sp>from<sp>unmapped<sp>List:<sp>{}" , name ) ; iterator . remove ( ) ; } } } catch ( java . lang . Exception ex ) { throw new org . apache . cassandra . hadoop . hive . metastore . CassandraHiveMetaStoreException ( "Could<sp>not<sp>retrieve<sp>unmapped<sp>keyspaces" , ex ) ; } return defs ; }
org . junit . Assert . assertTrue ( foundCreated )
testIsRepeatable ( ) { mockHelper . replayAll ( ) ; classUnderTest . setRepeatable ( true ) ; "<AssertPlaceHolder>" ; } isRepeatable ( ) { return false ; }
org . junit . Assert . assertEquals ( true , classUnderTest . isRepeatable ( ) )
testCopyListingFileStatusSerialization ( ) { org . apache . hadoop . tools . CopyListingFileStatus src = new org . apache . hadoop . tools . CopyListingFileStatus ( 4344L , false , 2 , ( 512 << 20 ) , 1234L , 5678L , new org . apache . hadoop . fs . permission . FsPermission ( ( ( short ) ( 330 ) ) ) , "dingo" , "yaks" , new org . apache . hadoop . fs . Path ( "hdfs://localhost:4344" ) ) ; org . apache . hadoop . io . DataOutputBuffer dob = new org . apache . hadoop . io . DataOutputBuffer ( ) ; src . write ( dob ) ; org . apache . hadoop . io . DataInputBuffer dib = new org . apache . hadoop . io . DataInputBuffer ( ) ; dib . reset ( dob . getData ( ) , 0 , dob . getLength ( ) ) ; org . apache . hadoop . tools . CopyListingFileStatus dst = new org . apache . hadoop . tools . CopyListingFileStatus ( ) ; dst . readFields ( dib ) ; "<AssertPlaceHolder>" ; } readFields ( java . io . DataInput ) { org . apache . hadoop . ozone . protocol . proto . OzoneManagerProtocolProtos . SecretKeyProto secretKey = org . apache . hadoop . ozone . protocol . proto . OzoneManagerProtocolProtos . SecretKeyProto . parseFrom ( ( ( java . io . DataInputStream ) ( in ) ) ) ; expiryDate = secretKey . getExpiryDate ( ) ; keyId = secretKey . getKeyId ( ) ; privateKey = org . apache . hadoop . hdds . security . x509 . keys . SecurityUtil . getPrivateKey ( secretKey . getPrivateKeyBytes ( ) . toByteArray ( ) , securityConfig ) ; publicKey = org . apache . hadoop . hdds . security . x509 . keys . SecurityUtil . getPublicKey ( secretKey . getPublicKeyBytes ( ) . toByteArray ( ) , securityConfig ) ; }
org . junit . Assert . assertEquals ( src , dst )
testApply ( ) { int size = 1000 ; final java . util . List < org . opencb . biodata . models . variant . protobuf . VcfSliceProtos . VcfSlice > lst = new java . util . ArrayList ( ) ; org . opencb . commons . io . DataWriter < org . opencb . biodata . models . variant . protobuf . VcfSliceProtos . VcfSlice > collector = new org . opencb . commons . io . DataWriter < org . opencb . biodata . models . variant . protobuf . VcfSliceProtos . VcfSlice > ( ) { @ org . opencb . opencga . storage . hadoop . variant . archive . Override public boolean write ( java . util . List < org . opencb . biodata . models . variant . protobuf . VcfSliceProtos . VcfSlice > batch ) { return lst . addAll ( batch ) ; } @ org . opencb . opencga . storage . hadoop . variant . archive . Override public boolean write ( org . opencb . biodata . models . variant . protobuf . VcfSliceProtos . VcfSlice elem ) { return lst . add ( elem ) ; } } ; org . opencb . commons . run . ParallelTaskRunner < org . opencb . biodata . models . variant . Variant , org . opencb . biodata . models . variant . protobuf . VcfSliceProtos . VcfSlice > parallelRunner = createParallelRunner ( size , collector ) ; parallelRunner . run ( ) ; "<AssertPlaceHolder>" ; } valueOf ( java . lang . String ) { return org . opencb . opencga . storage . core . variant . adaptors . VariantQueryParam . VALUES_MAP . get ( param ) ; }
org . junit . Assert . assertEquals ( java . lang . Integer . valueOf ( 2 ) , java . lang . Integer . valueOf ( lst . size ( ) ) )
testGetFileLength ( ) { org . apache . storm . blobstore . LocalFsBlobStore store = initLocalFs ( ) ; try ( org . apache . storm . blobstore . AtomicOutputStream out = store . createBlob ( "test" , new org . apache . storm . generated . SettableBlobMeta ( BlobStoreAclHandler . WORLD_EVERYTHING ) , null ) ) { out . write ( 1 ) ; } try ( org . apache . storm . blobstore . InputStreamWithMeta blobInputStream = store . getBlob ( "test" , null ) ) { "<AssertPlaceHolder>" ; } } getFileLength ( ) { return _path . length ( ) ; }
org . junit . Assert . assertEquals ( 1 , blobInputStream . getFileLength ( ) )
testNullContentTypeIgnored ( ) { final byte [ ] is = getBytesForCapabilities ( "DES3" ) ; final org . jscep . transport . response . Capabilities caps = fixture . getResponse ( is , null ) ; "<AssertPlaceHolder>" ; } getStrongestCipher ( ) { final java . lang . String cipher ; if ( ( cipherExists ( "AES" ) ) && ( caps . contains ( Capability . AES ) ) ) { cipher = "AES" ; } else if ( ( cipherExists ( "DESede" ) ) && ( caps . contains ( Capability . TRIPLE_DES ) ) ) { cipher = "DESede" ; } else { cipher = "DES" ; } return cipher ; }
org . junit . Assert . assertEquals ( "DESede" , caps . getStrongestCipher ( ) )
iterateImageSumStream1DotParallel ( ) { final org . eclipse . january . dataset . ILazyDataset lz = org . eclipse . january . dataset . DatasetFactory . ones ( 64 , 64 , 100 , 100 ) ; java . util . List < java . lang . Number > maxes = org . eclipse . dawnsci . analysis . dataset . SliceStreamSupport . sliceStream ( lz , 2 , 3 ) . parallel ( ) . map ( ( set ) -> set . max ( ) ) . collect ( java . util . stream . Collectors . toList ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return nameDataMappings . size ( ) ; }
org . junit . Assert . assertEquals ( ( 64 * 64 ) , maxes . size ( ) )
testBericht ( ) { final nl . bzk . migratiebrp . bericht . model . sync . impl . AdHocZoekPersoonAntwoordBericht bericht = org . mockito . Mockito . mock ( nl . bzk . migratiebrp . bericht . model . sync . impl . AdHocZoekPersoonAntwoordBericht . class ) ; final nl . bzk . migratiebrp . synchronisatie . runtime . service . adhoczoeken . AdHocZoekAntwoordNaarIscMessageCreator creator = new nl . bzk . migratiebrp . synchronisatie . runtime . service . adhoczoeken . AdHocZoekAntwoordNaarIscMessageCreator ( bericht ) ; org . mockito . Mockito . doThrow ( nl . bzk . migratiebrp . bericht . model . BerichtInhoudException . class ) . when ( bericht ) . format ( ) ; "<AssertPlaceHolder>" ; } createMessage ( javax . jms . Session ) { return receivedMessage ; }
org . junit . Assert . assertNull ( creator . createMessage ( session ) )
testAlertParameteredByOutDetail ( ) { logbookParameters . putParameterValue ( LogbookParameterName . outcomeDetail , ( ( ( eventType ) + "." ) + ( outcome ) ) ) ; logbookEvent . setOutDetail ( ( ( ( eventType ) + "." ) + ( outcome ) ) ) ; boolean isAlertEvent = alertLogbookOperationsDecorator . isAlertEvent ( logbookParameters ) ; "<AssertPlaceHolder>" ; } isAlertEvent ( fr . gouv . vitam . logbook . common . parameters . LogbookOperationParameters ) { for ( fr . gouv . vitam . common . model . logbook . LogbookEvent logbookEvent : alertEvents ) { if ( ( logbookEvent . getOutDetail ( ) ) != null ) { java . lang . String outDetail = parameters . getParameterValue ( LogbookParameterName . outcomeDetail ) ; if ( com . google . common . base . Strings . isNullOrEmpty ( outDetail ) ) { return false ; } if ( outDetail . equals ( logbookEvent . getOutDetail ( ) ) ) { return true ; } } else { java . lang . String eventType = parameters . getParameterValue ( LogbookParameterName . eventType ) ; if ( com . google . common . base . Strings . isNullOrEmpty ( eventType ) ) { return false ; } java . lang . String outcome = parameters . getParameterValue ( LogbookParameterName . outcome ) ; if ( com . google . common . base . Strings . isNullOrEmpty ( outcome ) ) { return false ; } if ( ( eventType . equals ( logbookEvent . getEvType ( ) ) ) && ( outcome . equals ( logbookEvent . getOutcome ( ) ) ) ) { return true ; } } } return false ; }
org . junit . Assert . assertTrue ( isAlertEvent )
putObjectList ( ) { com . alipay . simplehbase . myrecord . MyRecord [ ] records = mockSlims ( 4 ) ; java . util . List < com . alipay . simplehbase . client . PutRequest < com . alipay . simplehbase . myrecord . MyRecord > > requestList = new java . util . ArrayList < com . alipay . simplehbase . client . PutRequest < com . alipay . simplehbase . myrecord . MyRecord > > ( ) ; for ( com . alipay . simplehbase . myrecord . MyRecord myRecord : records ) { requestList . add ( new com . alipay . simplehbase . client . PutRequest < com . alipay . simplehbase . myrecord . MyRecord > ( myRecord . rowKey ( ) , myRecord ) ) ; } simpleHbaseClient . putObjectList ( requestList ) ; java . util . List < com . alipay . simplehbase . myrecord . MyRecord > resultList = simpleHbaseClient . findObjectList ( new com . alipay . simplehbase . myrecord . MyRecordRowKey ( 0 ) , new com . alipay . simplehbase . myrecord . MyRecordRowKey ( records . length ) , com . alipay . simplehbase . myrecord . MyRecord . class ) ; for ( int i = 0 ; i < ( requestList . size ( ) ) ; i ++ ) { "<AssertPlaceHolder>" ; } } get ( java . util . List ) { return table . get ( gets ) ; }
org . junit . Assert . assertEquals ( records [ i ] , resultList . get ( i ) )
cascadedEnum ( mockit . CascadingParametersTest$Foo ) { new mockit . Expectations ( ) { { mock . getBar ( ) . getEnum ( ) ; result = mockit . CascadingParametersTest . AnEnum . Second ; } } ; "<AssertPlaceHolder>" ; } getBar ( ) { return null ; }
org . junit . Assert . assertEquals ( mockit . CascadingParametersTest . AnEnum . Second , mock . getBar ( ) . getEnum ( ) )
getSelectedTokenNull ( ) { final org . obeonetwork . m2doc . services . TokenRegistry registry = new org . obeonetwork . m2doc . services . TokenRegistry ( ) ; "<AssertPlaceHolder>" ; } getSelectedToken ( org . obeonetwork . m2doc . properties . TemplateCustomProperties ) { final java . util . List < java . lang . String > res = new java . util . ArrayList ( ) ; for ( java . lang . String tokenName : getRegisteredTokens ( ) ) { boolean isSelected = true ; for ( java . util . Map . Entry < java . lang . String , java . util . List < java . lang . String > > entry : getServices ( tokenName ) . entrySet ( ) ) { final java . lang . String bundleName = entry . getKey ( ) ; for ( java . lang . String className : entry . getValue ( ) ) { if ( ! ( bundleName . equals ( customProperties . getServiceClasses ( ) . get ( className ) ) ) ) { isSelected = false ; break ; } } } final java . util . Set < java . lang . String > packages = new java . util . HashSet ( customProperties . getPackagesURIs ( ) ) ; for ( java . lang . String pkg : getPackages ( tokenName ) ) { if ( ! ( packages . contains ( pkg ) ) ) { isSelected = false ; break ; } } if ( isSelected ) { res . add ( tokenName ) ; } } return res ; }
org . junit . Assert . assertTrue ( registry . getSelectedToken ( null ) . isEmpty ( ) )
testInVMURI ( ) { locator . close ( ) ; org . apache . activemq . artemis . api . core . client . ServerLocator locator = addServerLocator ( org . apache . activemq . artemis . core . client . impl . ServerLocatorImpl . newLocator ( "vm:/1" ) ) ; org . apache . activemq . artemis . api . core . client . ClientSessionFactory factory = locator . createSessionFactory ( ) ; org . apache . activemq . artemis . api . core . client . ClientSession session = factory . createSession ( ) ; org . apache . activemq . artemis . api . core . client . ClientProducer producer = session . createProducer ( QUEUE ) ; producer . send ( session . createMessage ( true ) ) ; org . apache . activemq . artemis . api . core . client . ClientConsumer consumer = session . createConsumer ( QUEUE ) ; session . start ( ) ; "<AssertPlaceHolder>" ; session . close ( ) ; factory . close ( ) ; } receiveImmediate ( ) { return receive ( 0 , true ) ; }
org . junit . Assert . assertNotNull ( consumer . receiveImmediate ( ) )
notEqualToNull ( ) { com . mpatric . mp3agic . ID3v2TextFrameData frameData = new com . mpatric . mp3agic . ID3v2TextFrameData ( false , new com . mpatric . mp3agic . EncodedText ( EncodedText . TEXT_ENCODING_ISO_8859_1 , com . mpatric . mp3agic . ID3v2TextFrameDataTest . TEST_TEXT ) ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) return true ; if ( ! ( super . equals ( obj ) ) ) return false ; if ( ( getClass ( ) ) != ( obj . getClass ( ) ) ) return false ; com . mpatric . mp3agic . ID3v2CommentFrameData other = ( ( com . mpatric . mp3agic . ID3v2CommentFrameData ) ( obj ) ) ; if ( ( comment ) == null ) { if ( ( other . comment ) != null ) return false ; } else if ( ! ( comment . equals ( other . comment ) ) ) return false ; if ( ( description ) == null ) { if ( ( other . description ) != null ) return false ; } else if ( ! ( description . equals ( other . description ) ) ) return false ; if ( ( language ) == null ) { if ( ( other . language ) != null ) return false ; } else if ( ! ( language . equals ( other . language ) ) ) return false ; return true ; }
org . junit . Assert . assertFalse ( frameData . equals ( null ) )
testToStringNoQuota ( ) { org . apache . hadoop . fs . QuotaUsage quotaUsage = new org . apache . hadoop . fs . QuotaUsage . Builder ( ) . fileAndDirectoryCount ( 1234 ) . build ( ) ; java . lang . String expected = "<sp>none<sp>inf<sp>none" + "<sp>inf<sp>" ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( ( ( ( ( ( "OzoneFileSystem{URI=" + ( uri ) ) + ",<sp>" ) + "workingDir=" ) + ( workingDir ) ) + ",<sp>" ) + "userName=" ) + ( userName ) ) + ",<sp>" ) + "statistics=" ) + ( statistics ) ) + "}" ; }
org . junit . Assert . assertEquals ( expected , quotaUsage . toString ( ) )
testAddsTransistionListener ( ) { com . eclipsesource . tabris . internal . ui . UIDescriptor uiDescriptor = new com . eclipsesource . tabris . internal . ui . UIDescriptor ( ) ; com . eclipsesource . tabris . ui . TransitionListener listener = mock ( com . eclipsesource . tabris . ui . TransitionListener . class ) ; uiDescriptor . addTransitionListener ( listener ) ; java . util . List < com . eclipsesource . tabris . ui . TransitionListener > transitionListeners = uiDescriptor . getTransitionListeners ( ) ; "<AssertPlaceHolder>" ; } getTransitionListeners ( ) { return transitionListeners ; }
org . junit . Assert . assertTrue ( transitionListeners . contains ( listener ) )
permsShouldNotAffectOwners ( ) { java . util . List < org . candlepin . auth . permissions . Permission > perms = new java . util . ArrayList ( ) ; perms . add ( mock ( org . candlepin . auth . permissions . Permission . class ) ) ; org . candlepin . auth . UserPrincipal up = new org . candlepin . auth . UserPrincipal ( "admin" , perms , false ) ; "<AssertPlaceHolder>" ; } getOwners ( ) { java . util . List < org . candlepin . model . Owner > owners = new java . util . LinkedList ( ) ; for ( org . candlepin . auth . permissions . Permission permission : permissions ) { org . candlepin . model . Owner o = permission . getOwner ( ) ; if ( o != null ) { owners . add ( o ) ; } } return owners ; }
org . junit . Assert . assertTrue ( up . getOwners ( ) . isEmpty ( ) )
shouldFindWindowStores ( ) { mockThread ( true ) ; final java . util . List < org . apache . kafka . streams . state . ReadOnlyWindowStore < java . lang . Object , java . lang . Object > > windowStores = provider . stores ( "window-store" , org . apache . kafka . streams . state . QueryableStoreTypes . windowStore ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return ( headers ) == null ? 0 : headers . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , windowStores . size ( ) )
shouldReturnAnAbsentOptionalIfNoTargetDeviceCommandLineOptionsSet ( ) { com . facebook . buck . cli . TargetDeviceCommandLineOptions options = buildOptions ( ) ; "<AssertPlaceHolder>" ; } getTargetDeviceOptional ( ) { if ( ( ( ! ( getSerialNumber ( ) . isPresent ( ) ) ) && ( ! ( isEmulatorsOnlyModeEnabled ( ) ) ) ) && ( ! ( isRealDevicesOnlyModeEnabled ( ) ) ) ) { return java . util . Optional . empty ( ) ; } com . facebook . buck . android . device . TargetDevice device = new com . facebook . buck . android . device . TargetDevice ( ( isEmulatorsOnlyModeEnabled ( ) ? TargetDevice . Type . EMULATOR : TargetDevice . Type . REAL_DEVICE ) , getSerialNumber ( ) ) ; return java . util . Optional . of ( device ) ; }
org . junit . Assert . assertFalse ( options . getTargetDeviceOptional ( ) . isPresent ( ) )
testViewActionBitwiseValue ( ) { com . liferay . portal . kernel . model . ResourceAction viewResourceAction = com . liferay . portal . kernel . service . ResourceActionLocalServiceUtil . getResourceAction ( com . liferay . portal . service . ResourceActionLocalServiceTest . _NAME , ActionKeys . VIEW ) ; "<AssertPlaceHolder>" ; } getBitwiseValue ( ) { return _bitwiseValue ; }
org . junit . Assert . assertEquals ( 1L , viewResourceAction . getBitwiseValue ( ) )
isBasicCriteriaElementFactoryConstructorWorking ( ) { final java . lang . reflect . Constructor < com . uaihebert . uaicriteria . base . element . BasicCriteriaElementsFactory > constructor = com . uaihebert . uaicriteria . base . element . BasicCriteriaElementsFactory . class . getDeclaredConstructor ( new java . lang . Class [ 0 ] ) ; constructor . setAccessible ( true ) ; final com . uaihebert . uaicriteria . base . element . BasicCriteriaElementsFactory object = constructor . newInstance ( new java . lang . Object [ 0 ] ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( object )
findAllVerliesNLNationaliteit ( ) { final java . util . Collection < nl . bzk . brp . model . algemeen . stamgegeven . kern . RedenVerliesNLNationaliteit > result = stamTabelRepository . findAllVerliesNLNationaliteit ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return elementen . isEmpty ( ) ; }
org . junit . Assert . assertFalse ( result . isEmpty ( ) )
testEquals ( ) { org . jfree . data . time . Day day1 = new org . jfree . data . time . Day ( 29 , org . jfree . chart . date . MonthConstants . MARCH , 2002 ) ; org . jfree . data . time . Day day2 = new org . jfree . data . time . Day ( 29 , org . jfree . chart . date . MonthConstants . MARCH , 2002 ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( obj == ( this ) ) { return true ; } if ( ! ( obj instanceof org . jfree . data . xy . YWithXInterval ) ) { return false ; } org . jfree . data . xy . YWithXInterval that = ( ( org . jfree . data . xy . YWithXInterval ) ( obj ) ) ; if ( ( this . y ) != ( that . y ) ) { return false ; } if ( ( this . xLow ) != ( that . xLow ) ) { return false ; } if ( ( this . xHigh ) != ( that . xHigh ) ) { return false ; } return true ; }
org . junit . Assert . assertTrue ( day1 . equals ( day2 ) )
testMessageWithMissingTimestamp ( ) { java . util . List < org . apache . metron . profiler . MessageRoute > routes = router . route ( messageOne , createConfig ( profileWithEventTime ) , context ) ; "<AssertPlaceHolder>" ; } size ( ) { int size = 0 ; for ( java . util . Map m : variableMappings ) { size += m . size ( ) ; } return size ; }
org . junit . Assert . assertEquals ( 0 , routes . size ( ) )
whenNoReplicaCountSpecified_canChangeIt ( ) { domain . setReplicaCount ( "cluster1" , 7 ) ; "<AssertPlaceHolder>" ; } getReplicaCount ( java . lang . String ) { return getReplicaCountFor ( getCluster ( clusterName ) ) ; }
org . junit . Assert . assertThat ( domain . getReplicaCount ( "cluster1" ) , org . hamcrest . Matchers . equalTo ( 7 ) )
testIdleCommUser ( ) { boolean fail = false ; try { com . github . rinde . rinsim . core . model . comm . CommModel . builder ( ) . build ( com . github . rinde . rinsim . core . model . comm . CommModelTest . fakeDependencies ( ) ) . register ( new com . github . rinde . rinsim . core . model . comm . CommModelTest . IdleCommUser ( ) ) ; } catch ( final java . lang . IllegalStateException e ) { fail = true ; } "<AssertPlaceHolder>" ; } register ( java . lang . Void ) { return false ; }
org . junit . Assert . assertTrue ( fail )
testIsHasMoreShardsFalse ( ) { when ( mockDescription . getLastEvaluatedShardId ( ) ) . thenReturn ( null ) ; "<AssertPlaceHolder>" ; } isHasMoreShards ( ) { return ( internalDescription . getLastEvaluatedShardId ( ) ) != null ; }
org . junit . Assert . assertFalse ( adapter . isHasMoreShards ( ) )