input stringlengths 28 18.7k | output stringlengths 39 1.69k |
|---|---|
order ( ) { net . encomendaz . services . monitoring . Monitoring monitoring ; java . util . Map < java . lang . String , net . encomendaz . services . monitoring . Monitoring > monitorings = new java . util . TreeMap < java . lang . String , net . encomendaz . services . monitoring . Monitoring > ( ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "b" , "2" ) ; monitorings . put ( ( ( ( monitoring . getClientId ( ) ) + "-" ) + ( monitoring . getTrackId ( ) ) ) , monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "c" , "1" ) ; monitorings . put ( ( ( ( monitoring . getClientId ( ) ) + "-" ) + ( monitoring . getTrackId ( ) ) ) , monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "b" , "1" ) ; monitorings . put ( ( ( ( monitoring . getClientId ( ) ) + "-" ) + ( monitoring . getTrackId ( ) ) ) , monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "a" , "1" ) ; monitorings . put ( ( ( ( monitoring . getClientId ( ) ) + "-" ) + ( monitoring . getTrackId ( ) ) ) , monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "c" , "2" ) ; monitorings . put ( ( ( ( monitoring . getClientId ( ) ) + "-" ) + ( monitoring . getTrackId ( ) ) ) , monitoring ) ; java . util . List < net . encomendaz . services . monitoring . Monitoring > expectedOrder = new java . util . ArrayList < net . encomendaz . services . monitoring . Monitoring > ( ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "a" , "1" ) ; expectedOrder . add ( monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "b" , "1" ) ; expectedOrder . add ( monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "b" , "2" ) ; expectedOrder . add ( monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "c" , "1" ) ; expectedOrder . add ( monitoring ) ; monitoring = new net . encomendaz . services . monitoring . Monitoring ( "c" , "2" ) ; expectedOrder . add ( monitoring ) ; "<AssertPlaceHolder>" ; } getTrackId ( ) { return trackId ; } | org . junit . Assert . assertArrayEquals ( expectedOrder . toArray ( new net . encomendaz . services . monitoring . Monitoring [ 0 ] ) , monitorings . values ( ) . toArray ( new net . encomendaz . services . monitoring . Monitoring [ 0 ] ) ) |
copyConstructor_worksOnSafeCopy ( ) { org . eclipse . rap . json . JsonArray copy = new org . eclipse . rap . json . JsonArray ( array ) ; array . add ( 23 ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( ( childCollections . size ( ) ) == 0 ) && ( ( cheatsheets . size ( ) ) == 0 ) ; } | org . junit . Assert . assertTrue ( copy . isEmpty ( ) ) |
testUpdateNoUpload ( ) { java . lang . String nodeId = "101" ; java . lang . String parentId = "202" ; java . lang . String fileHandleId = "123" ; org . sagebionetworks . repo . model . Node newNode = new org . sagebionetworks . repo . model . Node ( ) ; newNode . setName ( "testCreateNode" ) ; newNode . setNodeType ( EntityType . folder ) ; newNode . setFileHandleId ( fileHandleId ) ; newNode . setParentId ( parentId ) ; newNode . setETag ( startEtag ) ; when ( mockEntityBootstrapper . getChildAclSchemeForPath ( "/root" ) ) . thenReturn ( ACL_SCHEME . INHERIT_FROM_PARENT ) ; org . mockito . ArgumentCaptor < org . sagebionetworks . repo . model . Node > argument = org . mockito . ArgumentCaptor . forClass ( org . sagebionetworks . repo . model . Node . class ) ; newNode . setId ( nodeId ) ; when ( mockNodeDao . createNewNode ( argument . capture ( ) ) ) . thenReturn ( newNode ) ; org . sagebionetworks . repo . model . UserInfo userInfo = anonUserInfo ; "<AssertPlaceHolder>" ; newNode . setId ( nodeId ) ; org . sagebionetworks . repo . model . Node oldNode = mock ( org . sagebionetworks . repo . model . Node . class ) ; when ( oldNode . getParentId ( ) ) . thenReturn ( parentId ) ; when ( mockNodeDao . getNode ( "101" ) ) . thenReturn ( oldNode ) ; when ( mockNodeDao . getParentId ( nodeId ) ) . thenReturn ( parentId ) ; nodeManager . update ( userInfo , newNode ) ; when ( mockAuthManager . canAccess ( userInfo , parentId , ObjectType . ENTITY , ACCESS_TYPE . UPLOAD ) ) . thenReturn ( AuthorizationManagerUtil . ACCESS_DENIED ) ; try { nodeManager . update ( userInfo , newNode ) ; org . junit . Assert . fail ( "expected<sp>UnauthorizedException" ) ; } catch ( org . sagebionetworks . repo . model . UnauthorizedException e ) { } } createNewNode ( org . sagebionetworks . repo . model . Node , org . sagebionetworks . repo . model . UserInfo ) { newNode = createNode ( newNode , userInfo ) ; return newNode . getId ( ) ; } | org . junit . Assert . assertEquals ( nodeId , nodeManager . createNewNode ( newNode , userInfo ) ) |
testGetElementsByEdgeIdWithViewRestrictedByGroup ( ) { final uk . gov . gchq . gaffer . graph . Graph graph = uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . getGraph ( ) ; final uk . gov . gchq . gaffer . operation . impl . add . AddElements addElements = new uk . gov . gchq . gaffer . operation . impl . add . AddElements . Builder ( ) . input ( uk . gov . gchq . gaffer . mapstore . impl . GetElementsHandlerTest . getElements ( ) ) . build ( ) ; graph . execute ( addElements , new uk . gov . gchq . gaffer . user . User ( ) ) ; final uk . gov . gchq . gaffer . operation . impl . get . GetElements getElements = new uk . gov . gchq . gaffer . operation . impl . get . GetElements . Builder ( ) . input ( new uk . gov . gchq . gaffer . operation . data . EdgeSeed ( "A" , "B0" , true ) ) . view ( new uk . gov . gchq . gaffer . data . elementdefinition . view . View . Builder ( ) . edge ( GetAllElementsHandlerTest . BASIC_EDGE1 ) . build ( ) ) . build ( ) ; final uk . gov . gchq . gaffer . commonutil . iterable . CloseableIterable < ? extends uk . gov . gchq . gaffer . data . element . Element > results = graph . execute ( getElements , new uk . gov . gchq . gaffer . user . User ( ) ) ; final java . util . Set < uk . gov . gchq . gaffer . data . element . Element > resultsSet = new java . util . HashSet ( ) ; uk . gov . gchq . gaffer . commonutil . stream . Streams . toStream ( results ) . forEach ( resultsSet :: add ) ; final java . util . Set < uk . gov . gchq . gaffer . data . element . Element > expectedResults = new java . util . HashSet ( ) ; uk . gov . gchq . gaffer . mapstore . impl . GetElementsHandlerTest . getElements ( ) . stream ( ) . filter ( ( element ) -> element . getGroup ( ) . equals ( GetAllElementsHandlerTest . BASIC_EDGE1 ) ) . filter ( ( element ) -> { if ( element instanceof uk . gov . gchq . gaffer . data . element . Entity ) { return ( ( ( uk . gov . gchq . gaffer . data . element . Entity ) ( element ) ) . getVertex ( ) . equals ( "A" ) ) || ( ( ( uk . gov . gchq . gaffer . data . element . Entity ) ( element ) ) . getVertex ( ) . equals ( "B0" ) ) ; } else { final uk . gov . gchq . gaffer . data . element . Edge edge = ( ( uk . gov . gchq . gaffer . data . element . Edge ) ( element ) ) ; return ( edge . getSource ( ) . equals ( "A" ) ) && ( edge . getDestination ( ) . equals ( "B0" ) ) ; } } ) . forEach ( expectedResults :: add ) ; "<AssertPlaceHolder>" ; } forEach ( java . util . function . BiConsumer ) { delegateMap . forEach ( action ) ; } | org . junit . Assert . assertEquals ( expectedResults , resultsSet ) |
testFullConfiguration ( ) { java . io . File projectCopy = this . resources . getBasedir ( "unit" ) ; java . io . File testPom = new java . io . File ( projectCopy , "full-test-plugin-config.xml" ) ; org . junit . Assume . assumeNotNull ( "POM<sp>file<sp>should<sp>not<sp>be<sp>null." , testPom ) ; org . junit . Assume . assumeTrue ( "POM<sp>file<sp>should<sp>exist<sp>as<sp>file." , ( ( testPom . exists ( ) ) && ( testPom . isFile ( ) ) ) ) ; com . wakaleo . schemaspy . SchemaSpyReport mojo = ( ( com . wakaleo . schemaspy . SchemaSpyReport ) ( this . rule . lookupMojo ( "schemaspy" , testPom ) ) ) ; mojo . executeReport ( java . util . Locale . getDefault ( ) ) ; java . io . File generatedFile = new java . io . File ( "./target/reports/full-test/schemaspy/index.html" ) ; System . out . println ( ( "generatedFile<sp>=<sp>" + ( generatedFile . getAbsolutePath ( ) ) ) ) ; "<AssertPlaceHolder>" ; } exists ( ) { return ( version . toString ( ) ) != null ; } | org . junit . Assert . assertTrue ( generatedFile . exists ( ) ) |
shouldIterateSuccessfullyWithRemoval ( ) { for ( java . lang . String v : values ) { multimap . put ( keys [ 0 ] , v ) ; } for ( java . util . Iterator < java . lang . String > iter = multimap . get ( keys [ 0 ] ) . iterator ( ) ; iter . hasNext ( ) ; ) { iter . next ( ) ; iter . remove ( ) ; } "<AssertPlaceHolder>" ; } isEmpty ( ) { return values . isEmpty ( ) ; } | org . junit . Assert . assertTrue ( multimap . isEmpty ( ) ) |
DefaultBehaviourTest ( ) { java . lang . String svgFile = ( com . itextpdf . svg . processors . DefaultSvgProcessorIntegrationTest . sourceFolder ) + "RedCircle.svg" ; java . io . InputStream svg = new java . io . FileInputStream ( svgFile ) ; com . itextpdf . styledxmlparser . node . impl . jsoup . JsoupXmlParser xmlParser = new com . itextpdf . styledxmlparser . node . impl . jsoup . JsoupXmlParser ( ) ; com . itextpdf . styledxmlparser . node . IDocumentNode root = xmlParser . parse ( svg , null ) ; com . itextpdf . svg . renderers . IBranchSvgNodeRenderer actual = ( ( com . itextpdf . svg . renderers . IBranchSvgNodeRenderer ) ( new com . itextpdf . svg . processors . impl . DefaultSvgProcessor ( ) . process ( root ) . getRootRenderer ( ) ) ) ; com . itextpdf . svg . renderers . IBranchSvgNodeRenderer expected = new com . itextpdf . svg . renderers . impl . SvgTagSvgNodeRenderer ( ) ; com . itextpdf . svg . renderers . ISvgNodeRenderer expectedEllipse = new com . itextpdf . svg . renderers . impl . EllipseSvgNodeRenderer ( ) ; java . util . Map < java . lang . String , java . lang . String > expectedEllipseAttributes = new java . util . HashMap ( ) ; expectedEllipse . setAttributesAndStyles ( expectedEllipseAttributes ) ; expected . addChild ( expectedEllipse ) ; "<AssertPlaceHolder>" ; } getChildren ( ) { if ( ( children ) == null ) { children = new java . util . ArrayList ( 0 ) ; } return children ; } | org . junit . Assert . assertEquals ( expected . getChildren ( ) . size ( ) , actual . getChildren ( ) . size ( ) ) |
testAnrAlsZoekcriteria ( ) { final nl . bzk . brp . business . regels . context . BerichtRegelContext regelContext = new nl . bzk . brp . business . regels . context . BerichtRegelContext ( null , nl . bzk . brp . model . algemeen . stamgegeven . kern . SoortAdministratieveHandeling . GEEF_DETAILS_PERSOON , maakBericht ( nl . bzk . brp . business . regels . bevraging . BRBV0001Test . BSN_NIET_AANWEZIG , nl . bzk . brp . business . regels . bevraging . BRBV0001Test . ANR_AANWEZIG , nl . bzk . brp . business . regels . bevraging . BRBV0001Test . OBJECTSLEUTEL_NIET_AANWEZIG ) ) ; final java . util . List < nl . bzk . brp . model . basis . BerichtIdentificeerbaar > overtreders = brbv0001 . valideer ( regelContext ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , overtreders . size ( ) ) |
testDeliveryResultTypeGets ( ) { java . lang . String code = "" ; try { code = _setupTestDeliveryResultType ( false ) ; _checkDeliveryResultTypeIntoDb ( code ) ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; "<AssertPlaceHolder>" ; } return ; } _checkDeliveryResultTypeIntoDb ( java . lang . String ) { org . isf . dlvrrestype . model . DeliveryResultType foundDeliveryResultType ; foundDeliveryResultType = ( ( org . isf . dlvrrestype . model . DeliveryResultType ) ( org . isf . dlvrrestype . test . Tests . jpa . find ( org . isf . dlvrrestype . model . DeliveryResultType . class , code ) ) ) ; org . isf . dlvrrestype . test . Tests . testDeliveryResultType . check ( foundDeliveryResultType ) ; return ; } | org . junit . Assert . assertEquals ( true , false ) |
testCompositeData ( ) { org . spf4j . jmx . JMXBeanMapping get = conv . get ( javax . management . openmbean . CompositeData . class ) ; "<AssertPlaceHolder>" ; } get ( org . spf4j . base . Tag ) { java . lang . Object res = baggage . get ( key ) ; if ( res == null ) { if ( ( source ) != null ) { return source . get ( key ) ; } else { return null ; } } else { return ( ( T ) ( res ) ) ; } } | org . junit . Assert . assertNotNull ( get ) |
objectWithSchemaDep ( ) { org . everit . json . schema . loader . ObjectSchema actual = ( ( org . everit . json . schema . loader . ObjectSchema ) ( org . everit . json . schema . loader . SchemaLoader . load ( org . everit . json . schema . loader . ObjectSchemaLoaderTest . get ( "objectWithSchemaDep" ) ) ) ) ; "<AssertPlaceHolder>" ; } getSchemaDependencies ( ) { return schemaDependencies ; } | org . junit . Assert . assertEquals ( 1 , actual . getSchemaDependencies ( ) . size ( ) ) |
concurrentApplyForEachTest ( ) { bigQueue = new com . leansoft . bigqueue . BigQueueImpl ( testDir , "concurrentApplyForEachTest" , BigArrayImpl . MINIMUM_DATA_PAGE_SIZE ) ; final long N = 100000 ; java . lang . Thread publisher = new java . lang . Thread ( new java . lang . Runnable ( ) { private com . leansoft . bigqueue . Long item = 1L ; @ com . leansoft . bigqueue . Override public void run ( ) { for ( long i = 0 ; i < N ; i ++ ) try { bigQueue . enqueue ( item . toString ( ) . getBytes ( ) ) ; ( item ) ++ ; java . lang . Thread . yield ( ) ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; } } } ) ; java . lang . Thread subscriber = new java . lang . Thread ( new java . lang . Runnable ( ) { private long item = 0L ; @ com . leansoft . bigqueue . Override public void run ( ) { for ( long i = 0 ; i < N ; i ++ ) try { if ( ( bigQueue . size ( ) ) > 0 ) { byte [ ] bytes = bigQueue . dequeue ( ) ; java . lang . String str = new java . lang . String ( bytes ) ; long curr = java . lang . Long . parseLong ( str ) ; "<AssertPlaceHolder>" ; item = curr ; } java . lang . Thread . yield ( ) ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; } } } ) ; subscriber . start ( ) ; publisher . start ( ) ; for ( long i = 0 ; i < N ; i += N / 100 ) { com . leansoft . bigqueue . BigQueueUnitTest . DefaultItemIterator dii = new com . leansoft . bigqueue . BigQueueUnitTest . DefaultItemIterator ( ) ; bigQueue . applyForEach ( dii ) ; System . out . println ( ( ( ( "[" + ( dii . getCount ( ) ) ) + "]<sp>" ) + ( dii . toString ( ) ) ) ) ; java . lang . Thread . sleep ( 2 ) ; } publisher . join ( ) ; subscriber . join ( ) ; } dequeue ( ) { long queueFrontIndex = - 1L ; try { queueFrontWriteLock . lock ( ) ; if ( this . isEmpty ( ) ) { return null ; } queueFrontIndex = this . queueFrontIndex . get ( ) ; byte [ ] data = this . innerArray . get ( queueFrontIndex ) ; long nextQueueFrontIndex = queueFrontIndex ; if ( nextQueueFrontIndex == ( Long . MAX_VALUE ) ) { nextQueueFrontIndex = 0L ; } else { nextQueueFrontIndex ++ ; } this . queueFrontIndex . set ( nextQueueFrontIndex ) ; com . leansoft . bigqueue . page . IMappedPage queueFrontIndexPage = this . queueFrontIndexPageFactory . acquirePage ( com . leansoft . bigqueue . BigQueueImpl . QUEUE_FRONT_PAGE_INDEX ) ; java . nio . ByteBuffer queueFrontIndexBuffer = queueFrontIndexPage . getLocal ( 0 ) ; queueFrontIndexBuffer . putLong ( nextQueueFrontIndex ) ; queueFrontIndexPage . setDirty ( true ) ; return data ; } finally { queueFrontWriteLock . unlock ( ) ; } } | org . junit . Assert . assertEquals ( ( ( item ) + 1 ) , curr ) |
testStopRecording ( ) { org . netbeans . modules . editor . lib2 . actions . MacroRecording macroRecording = org . netbeans . modules . editor . lib2 . actions . MacroRecording . get ( ) ; "<AssertPlaceHolder>" ; } stopRecording ( ) { allowRecording = false ; } | org . junit . Assert . assertNull ( macroRecording . stopRecording ( ) ) |
testFlowWithNoWorkflowData ( ) { java . net . URI resource = getClass ( ) . getResource ( "/corrupted/no_workflow_data_flow.sl" ) . toURI ( ) ; io . cloudslang . lang . compiler . modeller . result . ExecutableModellingResult result = compiler . preCompileSource ( io . cloudslang . lang . compiler . SlangSource . fromFile ( resource ) ) ; "<AssertPlaceHolder>" ; exception . expect ( io . cloudslang . lang . compiler . RuntimeException . class ) ; exception . expectMessage ( ( "Error<sp>compiling<sp>no_workflow_data_flow.sl.<sp>Flow:<sp>no_workflow_data<sp>has<sp>" + "no<sp>workflow<sp>property" ) ) ; throw result . getErrors ( ) . get ( 0 ) ; } getErrors ( ) { return errors ; } | org . junit . Assert . assertTrue ( ( ( result . getErrors ( ) . size ( ) ) > 0 ) ) |
addWithTranError ( ) { try { eqlTranService . addWithTranError ( ) ; org . junit . Assert . fail ( ) ; } catch ( java . lang . ArithmeticException e ) { } "<AssertPlaceHolder>" ; } queryDataCount ( ) { return eqlTranEqler . queryRecordCounts ( ) ; } | org . junit . Assert . assertThat ( eqlTranService . queryDataCount ( ) , org . hamcrest . CoreMatchers . is ( 0 ) ) |
shouldRetrieveMetaData ( ) { java . sql . Connection connection = connectToRemoteRepository ( ) ; java . sql . DatabaseMetaData metadata = connection . getMetaData ( ) ; "<AssertPlaceHolder>" ; readTables ( metadata ) ; } getMetaData ( ) { try { return new org . modeshape . jca . JcrManagedConnectionMetaData ( mcf . getRepository ( ) , session ) ; } catch ( java . lang . Exception e ) { throw new javax . resource . ResourceException ( e ) ; } } | org . junit . Assert . assertNotNull ( metadata ) |
testToevoegenStapels ( ) { final nl . bzk . migratiebrp . synchronisatie . dal . domein . brp . kern . entity . EntiteitSleutel persoonNationaliteit1Sleutel = new nl . bzk . migratiebrp . synchronisatie . dal . domein . brp . kern . entity . EntiteitSleutel ( nl . bzk . migratiebrp . synchronisatie . dal . domein . brp . kern . entity . Persoon . class , "persoonNationaliteitSet" , null ) ; persoonNationaliteit1Sleutel . addSleuteldeel ( "nation" , testData . persoonNationaliteit1 . getNationaliteit ( ) . getId ( ) ) ; final nl . bzk . migratiebrp . synchronisatie . dal . domein . brp . kern . entity . EntiteitSleutel persoonNationaliteit2Sleutel = new nl . bzk . migratiebrp . synchronisatie . dal . domein . brp . kern . entity . EntiteitSleutel ( nl . bzk . migratiebrp . synchronisatie . dal . domein . brp . kern . entity . Persoon . class , "persoonNationaliteitSet" , null ) ; persoonNationaliteit1Sleutel . addSleuteldeel ( "nation" , testData . persoonNationaliteit2 . getNationaliteit ( ) . getId ( ) ) ; final nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . Verschil verschil1 = new nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . Verschil ( persoonNationaliteit1Sleutel , null , testData . persoonNationaliteit1 , nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . VerschilType . RIJ_TOEGEVOEGD , null , null ) ; final nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . Verschil verschil2 = new nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . Verschil ( persoonNationaliteit2Sleutel , null , testData . persoonNationaliteit2 , nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . VerschilType . RIJ_TOEGEVOEGD , null , null ) ; testData . verwerker . verwerkWijzigingen ( vergelijkerResultaat ( verschil1 , verschil2 ) , testData . persoon , testData . administratieveHandeling ) ; "<AssertPlaceHolder>" ; } getPersoonNationaliteitSet ( ) { return persoonNationaliteitSet ; } | org . junit . Assert . assertEquals ( testData . persoonNationaliteitSet , testData . persoon . getPersoonNationaliteitSet ( ) ) |
testValidSetQueryTimeout ( ) { try ( java . sql . Statement statement = getConnection ( ) . createStatement ( ) ) { statement . setQueryTimeout ( 1000 ) ; "<AssertPlaceHolder>" ; } } getQueryTimeout ( ) { throwIfClosed ( ) ; return super . getQueryTimeout ( ) ; } | org . junit . Assert . assertThat ( statement . getQueryTimeout ( ) , org . hamcrest . CoreMatchers . equalTo ( 1000 ) ) |
testSVM1vN ( ) { setKSession ( getModelSession ( org . kie . pmml . pmml_4_2 . predictive . models . SVMTest . source2 , org . kie . pmml . pmml_4_2 . predictive . models . SVMTest . VERBOSE ) ) ; setKbase ( getKSession ( ) . getKieBase ( ) ) ; org . kie . api . runtime . KieSession kSession = getKSession ( ) ; kSession . fireAllRules ( ) ; org . kie . api . definition . type . FactType ztype = kSession . getKieBase ( ) . getFactType ( org . kie . pmml . pmml_4_2 . predictive . models . SVMTest . packageName , "Z" ) ; "<AssertPlaceHolder>" ; kSession . getEntryPoint ( "in_X" ) . insert ( 0.0 ) ; kSession . getEntryPoint ( "in_Y" ) . insert ( 0.0 ) ; kSession . fireAllRules ( ) ; checkFirstDataFieldOfTypeStatus ( ztype , true , false , "SVMXORMODEL" , "no" ) ; checkFirstDataFieldOfTypeStatus ( kSession . getKieBase ( ) . getFactType ( org . kie . pmml . pmml_4_2 . predictive . models . SVMTest . packageName , "OutZ" ) , true , false , "SVMXORMODEL" , "no" ) ; checkFirstDataFieldOfTypeStatus ( kSession . getKieBase ( ) . getFactType ( org . kie . pmml . pmml_4_2 . predictive . models . SVMTest . packageName , "ProbZNo" ) , true , false , "SVMXORMODEL" , 0.7357588 ) ; checkGeneratedRules ( ) ; } getKieBase ( ) { return null ; } | org . junit . Assert . assertNotNull ( ztype ) |
testGraphDependency ( ) { java . util . List < java . lang . String > expectedOrder = new java . util . ArrayList ( ) ; expectedOrder . add ( "A" ) ; expectedOrder . add ( "I" ) ; expectedOrder . add ( "B" ) ; expectedOrder . add ( "C" ) ; expectedOrder . add ( "H" ) ; expectedOrder . add ( "D" ) ; expectedOrder . add ( "G" ) ; expectedOrder . add ( "E" ) ; expectedOrder . add ( "F" ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphA = createGraph ( "A" ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphI = createGraph ( "I" ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphB = createGraph ( "B" ) ; graphB . addDependencyGraph ( graphA ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphC = createGraph ( "C" ) ; graphC . addDependencyGraph ( graphA ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphH = createGraph ( "H" ) ; graphH . addDependencyGraph ( graphI ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphG = createGraph ( "G" ) ; graphG . addDependencyGraph ( graphC ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphE = createGraph ( "E" ) ; graphE . addDependencyGraph ( graphB ) ; graphE . addDependencyGraph ( graphG ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphD = createGraph ( "D" ) ; graphD . addDependencyGraph ( graphB ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphF = createGraph ( "F" ) ; graphF . addDependencyGraph ( graphD ) ; graphF . addDependencyGraph ( graphE ) ; graphF . addDependencyGraph ( graphH ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > dag = graphF ; dag . prepareForEnumeration ( ) ; com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder nextNode = dag . getNext ( ) ; int i = 0 ; while ( nextNode != null ) { "<AssertPlaceHolder>" ; dag . reportCompletion ( nextNode ) ; nextNode = dag . getNext ( ) ; i ++ ; } } get ( java . lang . String ) { return this . getAsync ( webhookName ) . toBlocking ( ) . single ( ) ; } | org . junit . Assert . assertEquals ( expectedOrder . get ( i ) , nextNode . key ( ) ) |
testEqualsSameReference ( ) { io . joynr . messaging . info . PerformanceMeasures m = new io . joynr . messaging . info . PerformanceMeasures ( ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ! ( arg0 instanceof io . joynr . dispatching . subscription . PublicationManagerImpl . PublicationInformation ) ) { return false ; } io . joynr . dispatching . subscription . PublicationManagerImpl . PublicationInformation pi = ( ( io . joynr . dispatching . subscription . PublicationManagerImpl . PublicationInformation ) ( arg0 ) ) ; return ( ( proxyParticipantId . equals ( pi . proxyParticipantId ) ) && ( providerParticipantId . equals ( pi . providerParticipantId ) ) ) && ( subscriptionRequest . equals ( pi . subscriptionRequest ) ) ; } | org . junit . Assert . assertTrue ( m . equals ( m ) ) |
testFeatures ( ) { "<AssertPlaceHolder>" ; } features ( ) { return graph . features ( ) ; } | org . junit . Assert . assertEquals ( g . features ( ) . toString ( ) , graph . features ( ) . toString ( ) ) |
testMultipleInvocationSingleThread ( ) { final int testLeft1 = 13 ; final int testRight1 = 23 ; final int testLeft2 = 14 ; final int testRight2 = 1001 ; final int expected1 = testLeft1 + ( 31 * testRight1 ) ; final int expected2 = testLeft2 + ( 31 * testRight2 ) ; final java . util . concurrent . atomic . AtomicInteger i = new java . util . concurrent . atomic . AtomicInteger ( 0 ) ; final org . apache . reef . wake . impl . MergingEventHandler < java . lang . Integer , java . lang . Integer > dut = new org . apache . reef . wake . impl . MergingEventHandler ( new org . apache . reef . wake . EventHandler < org . apache . reef . wake . impl . MergingEventHandler . Pair < java . lang . Integer , java . lang . Integer > > ( ) { @ org . apache . reef . wake . test . Override public void onNext ( final org . apache . reef . wake . impl . MergingEventHandler . Pair < java . lang . Integer , java . lang . Integer > value ) { i . addAndGet ( ( ( value . getFirst ( ) ) + ( 31 * ( value . getSecond ( ) ) ) ) ) ; } } ) ; dut . getLeft ( ) . onNext ( testLeft1 ) ; dut . getRight ( ) . onNext ( testRight1 ) ; dut . getLeft ( ) . onNext ( testLeft2 ) ; dut . getRight ( ) . onNext ( testRight2 ) ; "<AssertPlaceHolder>" ; } get ( ) { if ( ( instance ) != null ) { return instance ; } try { synchronized ( injector ) { final T t ; if ( org . apache . reef . tang . annotations . Name . class . isAssignableFrom ( iface ) ) { t = injector . getNamedInstance ( ( ( java . lang . Class < org . apache . reef . tang . annotations . Name < T > > ) ( iface ) ) ) ; } else { t = injector . getInstance ( iface ) ; } final org . apache . reef . tang . Aspect a = injector . getAspect ( ) ; if ( a != null ) { a . injectionFutureInstantiated ( this , t ) ; } return t ; } } catch ( final org . apache . reef . tang . exceptions . InjectionException e ) { throw new java . lang . RuntimeException ( e ) ; } } | org . junit . Assert . assertEquals ( ( expected1 + expected2 ) , i . get ( ) ) |
testToList ( ) { org . antlr . v4 . runtime . misc . IntervalSet s = org . antlr . v4 . runtime . misc . IntervalSet . of ( 20 , 25 ) ; s . add ( 50 , 55 ) ; s . add ( 5 , 5 ) ; java . lang . String expecting = "[5,<sp>20,<sp>21,<sp>22,<sp>23,<sp>24,<sp>25,<sp>50,<sp>51,<sp>52,<sp>53,<sp>54,<sp>55]" ; java . lang . String result = java . lang . String . valueOf ( s . toList ( ) ) ; "<AssertPlaceHolder>" ; } toList ( ) { java . util . List < java . lang . Integer > values = new java . util . ArrayList < java . lang . Integer > ( ) ; int n = intervals . size ( ) ; for ( int i = 0 ; i < n ; i ++ ) { org . antlr . v4 . runtime . misc . Interval I = intervals . get ( i ) ; int a = I . a ; int b = I . b ; for ( int v = a ; v <= b ; v ++ ) { values . add ( v ) ; } } return values ; } | org . junit . Assert . assertEquals ( expecting , result ) |
testSeed ( ) { info . debatty . java . lsh . MinHash mh = new info . debatty . java . lsh . MinHash ( 100 , 100 , 123456 ) ; info . debatty . java . lsh . MinHash mh2 = new info . debatty . java . lsh . MinHash ( 100 , 100 , 123456 ) ; java . util . Random r = new java . util . Random ( ) ; java . util . Set < java . lang . Integer > ints = new java . util . HashSet < java . lang . Integer > ( ) ; for ( int i = 0 ; i < 50 ; i ++ ) { ints . add ( r . nextInt ( ) ) ; } "<AssertPlaceHolder>" ; } signature ( java . util . Set ) { int [ ] sig = new int [ n ] ; for ( int i = 0 ; i < ( n ) ; i ++ ) { sig [ i ] = Integer . MAX_VALUE ; } final java . util . List < java . lang . Integer > list = new java . util . ArrayList < java . lang . Integer > ( set ) ; java . util . Collections . sort ( list ) ; for ( final int r : list ) { for ( int i = 0 ; i < ( n ) ; i ++ ) { sig [ i ] = java . lang . Math . min ( sig [ i ] , h ( i , r ) ) ; } } return sig ; } | org . junit . Assert . assertArrayEquals ( mh . signature ( ints ) , mh2 . signature ( ints ) ) |
testSsh ( ) { io . cdap . cdap . common . ssh . SSHConfig config = getSSHConfig ( ) ; try ( io . cdap . cdap . runtime . spi . ssh . SSHSession session = new io . cdap . cdap . common . ssh . DefaultSSHSession ( config ) ) { for ( int i = 0 ; i < 10 ; i ++ ) { java . lang . String msg = "Sending<sp>some<sp>message<sp>" + i ; java . lang . String result = session . executeAndWait ( msg ) ; "<AssertPlaceHolder>" ; } } try ( io . cdap . cdap . runtime . spi . ssh . SSHSession session = new io . cdap . cdap . common . ssh . DefaultSSHSession ( config ) ) { try { session . executeAndWait ( "failure" ) ; org . junit . Assert . fail ( "Expected<sp>failure<sp>from<sp>ssh<sp>command" ) ; } catch ( java . lang . Exception e ) { } } } executeAndWait ( java . util . List ) { return getDelegate ( ) . executeAndWait ( commands ) ; } | org . junit . Assert . assertEquals ( msg , result ) |
asyncPut_A$String$byteArray$String ( ) { com . m3 . curly . com . m3 . curly . Response response = com . m3 . curly . com . m3 . curly . HTTP . asyncPut ( "http://localhost:8808" , "<user><id>1234</id><name>Andy</name></user>" . getBytes ( ) , "text/xml" ) . get ( ) ; "<AssertPlaceHolder>" ; } getStatus ( ) { return status ; } | org . junit . Assert . assertThat ( response . getStatus ( ) , is ( 200 ) ) |
testGetLabelWheNameIsValidAndTitleIsValid ( ) { final java . lang . String title = "title" ; final java . lang . String name = "name" ; doReturn ( name ) . when ( factory ) . getName ( element ) ; doReturn ( title ) . when ( factory ) . getTitle ( element ) ; final java . lang . String label = factory . getLabel ( element ) ; "<AssertPlaceHolder>" ; } getLabel ( org . kie . workbench . common . stunner . core . graph . Node ) { final java . util . Optional < org . kie . workbench . common . dmn . api . definition . v1_1 . Expression > expression = java . util . Optional . of ( getExpression ( node ) ) ; return expressionEditorDefinitionsSupplier . get ( ) . getExpressionEditorDefinition ( expression ) . get ( ) . getName ( ) ; } | org . junit . Assert . assertEquals ( label , name ) |
shouldProcessesDispatcherThatReturnsNull ( ) { final uk . gov . justice . services . messaging . JsonEnvelope inputEnvelope = mock ( uk . gov . justice . services . messaging . JsonEnvelope . class ) ; final java . lang . String component = "component" ; when ( interceptorCache . getInterceptors ( component ) ) . thenReturn ( interceptors ( ) ) ; when ( dispatch . apply ( inputEnvelope ) ) . thenReturn ( null ) ; final uk . gov . justice . services . core . interceptor . DefaultInterceptorChainProcessor interceptorChainProcessor = new uk . gov . justice . services . core . interceptor . DefaultInterceptorChainProcessor ( interceptorCache , dispatch , component ) ; final java . util . Optional < uk . gov . justice . services . messaging . JsonEnvelope > result = interceptorChainProcessor . process ( inputEnvelope ) ; "<AssertPlaceHolder>" ; } process ( uk . gov . justice . services . messaging . JsonEnvelope ) { final uk . gov . justice . services . core . interceptor . InterceptorContext context = uk . gov . justice . services . core . interceptor . InterceptorContext . interceptorContextWithInput ( jsonEnvelope ) ; context . setInputParameter ( "component" , component ) ; return process ( context ) ; } | org . junit . Assert . assertThat ( result , org . hamcrest . core . Is . is ( java . util . Optional . empty ( ) ) ) |
testBroadcasts ( ) { org . qcri . rheem . tests . Collection < java . lang . Integer > broadcastedValues = org . qcri . rheem . tests . Arrays . asList ( 1 , 2 , 3 , 4 ) ; org . qcri . rheem . tests . Collection < java . lang . Integer > mainValues = org . qcri . rheem . tests . Arrays . asList ( 2 , 4 , 6 , 2 ) ; org . qcri . rheem . tests . List < java . lang . Integer > collectedValues = new org . qcri . rheem . tests . ArrayList ( ) ; org . qcri . rheem . tests . List < java . lang . Integer > expectedValues = org . qcri . rheem . tests . Arrays . asList ( 2 , 2 , 4 ) ; final org . qcri . rheem . core . types . DataSetType < java . lang . Integer > integerDataSetType = org . qcri . rheem . core . types . DataSetType . createDefault ( org . qcri . rheem . tests . Integer . class ) ; org . qcri . rheem . basic . operators . CollectionSource < java . lang . Integer > broadcastSource = new org . qcri . rheem . basic . operators . CollectionSource ( broadcastedValues , integerDataSetType ) ; org . qcri . rheem . basic . operators . CollectionSource < java . lang . Integer > mainSource = new org . qcri . rheem . basic . operators . CollectionSource ( mainValues , integerDataSetType ) ; org . qcri . rheem . basic . operators . FilterOperator < java . lang . Integer > semijoin = new org . qcri . rheem . basic . operators . FilterOperator ( integerDataSetType , new org . qcri . rheem . core . function . PredicateDescriptor . ExtendedSerializablePredicate < java . lang . Integer > ( ) { private Set < java . lang . Integer > allowedInts ; @ java . lang . Override public void open ( org . qcri . rheem . core . function . ExecutionContext ctx ) { this . allowedInts = new org . qcri . rheem . tests . HashSet < > ( ctx . < java . lang . Integer > getBroadcast ( "allowed<sp>values" ) ) ; } @ java . lang . Override public boolean test ( java . lang . Integer integer ) { return this . allowedInts . contains ( integer ) ; } } ) ; final org . qcri . rheem . basic . operators . LocalCallbackSink < java . lang . Integer > collectingSink = org . qcri . rheem . basic . operators . LocalCallbackSink . createCollectingSink ( collectedValues , integerDataSetType ) ; mainSource . connectTo ( 0 , semijoin , 0 ) ; broadcastSource . broadcastTo ( 0 , semijoin , "allowed<sp>values" ) ; semijoin . connectTo ( 0 , collectingSink , 0 ) ; org . qcri . rheem . core . plan . rheemplan . RheemPlan rheemPlan = new org . qcri . rheem . core . plan . rheemplan . RheemPlan ( collectingSink ) ; org . qcri . rheem . core . api . RheemContext rheemContext = new org . qcri . rheem . core . api . RheemContext ( ) . with ( org . qcri . rheem . java . Java . basicPlugin ( ) ) ; rheemContext . execute ( rheemPlan ) ; org . qcri . rheem . tests . Collections . sort ( collectedValues ) ; "<AssertPlaceHolder>" ; } execute ( org . qcri . rheem . core . platform . CrossPlatformExecutor$StageActivator ) { final org . qcri . rheem . core . platform . ExecutionStage stage = stageActivator . getStage ( ) ; final org . qcri . rheem . core . optimizer . OptimizationContext optimizationContext = stageActivator . getOptimizationContext ( ) ; this . instrumentationStrategy . applyTo ( stage ) ; org . qcri . rheem . core . platform . Executor executor = this . getOrCreateExecutorFor ( stage ) ; this . logger . info ( "Having<sp>{}<sp>execute<sp>{}:\n{}" , executor , stage , stage . getPlanAsString ( "><sp>" ) ) ; long startTime = java . lang . System . currentTimeMillis ( ) ; executor . execute ( stage , optimizationContext , this ) ; long finishTime = java . lang . System . currentTimeMillis ( ) ; this . logger . info ( "Executed<sp>{}<sp>in<sp>{}." , stage , org . qcri . rheem . core . util . Formats . formatDuration ( ( finishTime - startTime ) , true ) ) ; this . completedStages . add ( stage ) ; if ( stage . isLoopHead ( ) ) { this . getOrCreateLoopContext ( stage . getLoop ( ) ) . scrapPreviousTransitionContext ( ) ; } } | org . junit . Assert . assertEquals ( expectedValues , collectedValues ) |
testGetTableOptions ( ) { com . bazaarvoice . emodb . sor . api . TableOptions expected = new com . bazaarvoice . emodb . sor . api . TableOptionsBuilder ( ) . setPlacement ( "my:placement" ) . build ( ) ; when ( _server . getTableOptions ( "table-name" ) ) . thenReturn ( expected ) ; com . bazaarvoice . emodb . sor . api . TableOptions actual = blobClient ( ) . getTableOptions ( "table-name" ) ; "<AssertPlaceHolder>" ; verify ( _server ) . getTableOptions ( "table-name" ) ; verifyNoMoreInteractions ( _server ) ; } getTableOptions ( java . lang . String ) { return _dataStore . getTableOptions ( table ) ; } | org . junit . Assert . assertEquals ( actual , expected ) |
shouldReturnUnmodifiableComponentsWhenLocked ( ) { final uk . gov . gchq . gaffer . data . element . function . ElementAggregator aggregator = new uk . gov . gchq . gaffer . data . element . function . ElementAggregator ( ) ; aggregator . lock ( ) ; final java . util . List < uk . gov . gchq . koryphe . tuple . binaryoperator . TupleAdaptedBinaryOperator < java . lang . String , ? > > components = aggregator . getComponents ( ) ; try { components . add ( null ) ; org . junit . Assert . fail ( "Exception<sp>expected" ) ; } catch ( final java . lang . UnsupportedOperationException e ) { "<AssertPlaceHolder>" ; } } add ( uk . gov . gchq . gaffer . data . element . Element ) { if ( null == element ) { return ; } if ( null == ( queue ) ) { queue = new uk . gov . gchq . gaffer . commonutil . iterable . ConsumableBlockingQueue ( maxQueueSize ) ; restart = true ; } try { queue . put ( element ) ; } catch ( final java . lang . InterruptedException e ) { throw new java . lang . RuntimeException ( "Interrupted<sp>while<sp>waiting<sp>to<sp>add<sp>an<sp>element<sp>to<sp>the<sp>queue" , e ) ; } if ( ( restart ) && ( ! ( queue . isEmpty ( ) ) ) ) { restart = false ; store . runAsync ( ( ) -> { try { store . execute ( new uk . gov . gchq . gaffer . operation . impl . add . AddElements . Builder ( ) . input ( queue ) . validate ( validate ) . skipInvalidElements ( skipInvalid ) . build ( ) , new uk . gov . gchq . gaffer . store . Context ( new uk . gov . gchq . gaffer . user . User ( ) ) ) ; restart = true ; } catch ( final e ) { throw new < uk . gov . gchq . gaffer . flink . operation . handler . e > java . lang . RuntimeException ( uk . gov . gchq . gaffer . flink . operation . handler . e . getMessage ( ) ) ; } } ) ; } } | org . junit . Assert . assertNotNull ( e ) |
testNumberOfTransactions ( ) { java . lang . String name = "distrlog-txncount" ; org . apache . distributedlog . api . DistributedLogManager dlm = createNewDLM ( conf , name ) ; org . apache . distributedlog . BKSyncLogWriter out = ( ( org . apache . distributedlog . BKSyncLogWriter ) ( dlm . startLogSegmentNonPartitioned ( ) ) ) ; for ( long i = 1 ; i <= 100 ; i ++ ) { org . apache . distributedlog . LogRecord op = org . apache . distributedlog . DLMTestUtil . getLogRecordInstance ( i ) ; out . write ( op ) ; } out . closeAndComplete ( ) ; long numTrans = org . apache . distributedlog . DLMTestUtil . getNumberofLogRecords ( createNewDLM ( conf , name ) , 1 ) ; "<AssertPlaceHolder>" ; dlm . close ( ) ; } createNewDLM ( org . apache . distributedlog . DistributedLogConfiguration , java . lang . String ) { return createNewDLM ( conf , name , PermitLimiter . NULL_PERMIT_LIMITER ) ; } | org . junit . Assert . assertEquals ( 100 , numTrans ) |
testExtractEmptyFieldSet ( ) { org . apache . lucene . search . Query q = mock ( org . apache . lucene . search . Query . class ) ; com . sematext . lucene . query . extractor . GenericQueryExtractor genericQueryExtractor = new com . sematext . lucene . query . extractor . GenericQueryExtractor ( ) ; java . util . Set < java . lang . String > extractedFieldNames = new java . util . HashSet ( ) ; genericQueryExtractor . extractSubQueriesFields ( q , com . sematext . lucene . query . extractor . TestQueryExtractor . DEFAULT_EXTRACTORS , extractedFieldNames ) ; "<AssertPlaceHolder>" ; } extractSubQueriesFields ( org . apache . lucene . search . TermQuery , java . lang . Iterable , java . util . Set ) { extractedFields . add ( q . getTerm ( ) . field ( ) ) ; } | org . junit . Assert . assertEquals ( 0 , extractedFieldNames . size ( ) ) |
findFunctionsByNameInNamespaceForResourceWithSameNamedFunctionsOtherResourceNotInNamespaceShouldReturnEmptyList ( ) { ch . puzzle . itc . mobiliar . business . resourcegroup . entity . ResourceTypeEntity rootResourceType = createRootResourceType ( ) ; ch . puzzle . itc . mobiliar . business . resourcegroup . entity . ResourceTypeEntity subResourceType = createSubResourceType ( rootResourceType ) ; ch . puzzle . itc . mobiliar . business . resourcegroup . entity . ResourceTypeEntity otherSubResourceType = createSubResourceType ( rootResourceType ) ; ch . puzzle . itc . mobiliar . business . resourcegroup . entity . ResourceEntity resource = createResourceWithType ( "amw" , 1000 , subResourceType ) ; ch . puzzle . itc . mobiliar . business . resourcegroup . entity . ResourceEntity otherResource = createResourceWithType ( "amw2" , 1001 , otherSubResourceType , OTHER_FUNCTION_A ) ; when ( resourceRepositoryMock . loadWithFunctionsAndMiksForId ( resource . getId ( ) ) ) . thenReturn ( resource ) ; ch . puzzle . itc . mobiliar . business . function . control . List < ch . puzzle . itc . mobiliar . business . function . entity . AmwFunctionEntity > functionsWithName = functionService . findFunctionsByNameInNamespace ( resource , FUNCTION_A . getName ( ) ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { if ( ( ( asProperties ) != null ) && ( ! ( asProperties . isEmpty ( ) ) ) ) { return false ; } if ( ( ( nodeProperties ) != null ) && ( ! ( nodeProperties . isEmpty ( ) ) ) ) { return false ; } if ( ( ( consumerUnit ) != null ) && ( ! ( consumerUnit . isEmpty ( ) ) ) ) { return false ; } return true ; } | org . junit . Assert . assertTrue ( functionsWithName . isEmpty ( ) ) |
testPreviousValue ( ) { org . jacop . IntDomain goldenResultDomain = ( ( org . jacop . IntDomain ) ( prepareMethod . invoke ( this , new java . lang . Object [ ] { new int [ ] { 1 , 3 , 5 , 7 , 12 , 18 } } ) ) ) ; "<AssertPlaceHolder>" ; } previousValue ( int ) { for ( int m = ( size ) - 1 ; m >= 0 ; m -- ) { org . jacop . core . Interval i = intervals [ m ] ; if ( ( i . min ) >= value ) continue ; if ( ( ( i . max ) >= value ) && ( ( i . min ) < value ) ) return value - 1 ; if ( ( i . max ) < value ) return i . max ; } return value ; } | org . junit . Assert . assertEquals ( 1 , goldenResultDomain . previousValue ( 2 ) ) |
testDoubleTest2 ( ) { org . openl . rules . testmethod . TestDoubleDelta . ITestDouble instance = org . openl . rules . TestUtils . create ( org . openl . rules . testmethod . TestDoubleDelta . FILE_NAME , org . openl . rules . testmethod . TestDoubleDelta . ITestDouble . class ) ; org . openl . rules . testmethod . TestUnitsResults result = instance . geTestDoubleTest2 ( ) ; "<AssertPlaceHolder>" ; } getNumberOfFailures ( ) { int cnt = 0 ; for ( int i = 0 ; i < ( getNumberOfTestUnits ( ) ) ; i ++ ) { if ( ( testUnits . get ( i ) . getResultStatus ( ) ) != ( TestStatus . TR_OK ) ) { ++ cnt ; } } return cnt ; } | org . junit . Assert . assertEquals ( 1 , result . getNumberOfFailures ( ) ) |
getEncoderReturnsNullIfV2TagDoesNotExist ( ) { com . mpatric . mp3agic . ID3v1 id3v1Tag = new com . mpatric . mp3agic . ID3WrapperTest . ID3v1TagForTesting ( ) ; com . mpatric . mp3agic . ID3Wrapper wrapper = new com . mpatric . mp3agic . ID3Wrapper ( id3v1Tag , null ) ; "<AssertPlaceHolder>" ; } getEncoder ( ) { if ( ( id3v2Tag ) != null ) { return id3v2Tag . getEncoder ( ) ; } else { return null ; } } | org . junit . Assert . assertNull ( wrapper . getEncoder ( ) ) |
testGetJmsReplyToWithNullAddressWithConsumerDestReturnsNull ( ) { org . apache . qpid . jms . provider . amqp . message . AmqpJmsMessageFacade message = org . mockito . Mockito . mock ( org . apache . qpid . jms . provider . amqp . message . AmqpJmsMessageFacade . class ) ; org . mockito . Mockito . when ( message . getReplyToAddress ( ) ) . thenReturn ( null ) ; org . apache . qpid . jms . JmsQueue consumerDestination = new org . apache . qpid . jms . JmsQueue ( "ConsumerDestination" ) ; "<AssertPlaceHolder>" ; } getJmsReplyTo ( org . apache . qpid . jms . provider . amqp . message . AmqpJmsMessageFacade , org . apache . qpid . jms . JmsDestination ) { java . lang . String replyTo = message . getReplyToAddress ( ) ; byte typeByte = org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . getTypeByte ( message , org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . JMS_REPLY_TO_TYPE_MSG_ANNOTATION_SYMBOL ) ; if ( typeByte == ( org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . UNKNOWN_TYPE ) ) { typeByte = org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . getTypeByte ( message , AmqpMessageSupport . LEGACY_REPLY_TO_TYPE_MSG_ANNOTATION_SYMBOL ) ; } java . lang . String name = org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . stripPrefixIfNecessary ( replyTo , message . getConnection ( ) , typeByte , consumerDestination ) ; return org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . createDestination ( name , typeByte , consumerDestination , true ) ; } | org . junit . Assert . assertNull ( org . apache . qpid . jms . provider . amqp . message . AmqpDestinationHelper . getJmsReplyTo ( message , consumerDestination ) ) |
testMatchesExactElementsFirst ( ) { final com . allanbank . mongodb . bson . Document t = com . allanbank . mongodb . bson . builder . BuilderFactory . start ( ) . addInteger ( "a" , 1 ) . addInteger ( "b" , 2 ) . build ( ) ; final com . allanbank . mongodb . bson . Document [ ] tags = new com . allanbank . mongodb . bson . Document [ ] { com . allanbank . mongodb . bson . builder . BuilderFactory . start ( ) . addInteger ( "a" , 1 ) . addInteger ( "b" , 2 ) . build ( ) , com . allanbank . mongodb . bson . builder . BuilderFactory . start ( ) . addInteger ( "c" , 1 ) . build ( ) } ; for ( final com . allanbank . mongodb . ReadPreference . Mode mode : ReadPreference . Mode . values ( ) ) { final com . allanbank . mongodb . ReadPreference prefs = new com . allanbank . mongodb . ReadPreference ( mode , tags ) ; "<AssertPlaceHolder>" ; } } matches ( java . util . regex . Pattern ) { myEqualsComparison = null ; myOtherComparisons . put ( MiscellaneousOperator . REG_EX , new com . allanbank . mongodb . bson . element . RegularExpressionElement ( MiscellaneousOperator . REG_EX . getToken ( ) , pattern ) ) ; return this ; } | org . junit . Assert . assertTrue ( prefs . matches ( t ) ) |
allPosibilites ( ) { try { for ( int i = 0 ; i < ( getNumberOfPossibilities ( ) ) ; ++ i ) { abuseWeakness ( i , null , new wsattacker . library . signatureWrapping . option . PayloadElement ( wsattacker . library . signatureWrapping . xpath . weakness . XPathDescendantWeaknessAllPossibilitiesTest . payloadElement , null ) ) ; } } catch ( java . lang . Exception e ) { throw e ; } finally { org . apache . log4j . Logger log = org . apache . log4j . Logger . getLogger ( getClass ( ) ) ; java . lang . StringBuffer buf = new java . lang . StringBuffer ( ) ; buf . append ( "All<sp>Calls:\n" ) ; for ( int i = 0 ; i < ( callList . size ( ) ) ; ++ i ) { buf . append ( java . lang . String . format ( "%2d<sp>==><sp>%s\n" , i , callList . get ( i ) ) ) ; } log . info ( buf ) ; } "<AssertPlaceHolder>" ; } size ( ) { return options . size ( ) ; } | org . junit . Assert . assertEquals ( ( ( ( ( 2 * 3 ) + ( 2 * 1 ) ) + ( 2 * 2 ) ) * ( 3 + 3 ) ) , callList . size ( ) ) |
testDuplicateBind ( ) { bind ( false ) ; try { acceptor . bind ( ) ; org . junit . Assert . fail ( "Exception<sp>is<sp>not<sp>thrown" ) ; } catch ( java . lang . Exception e ) { "<AssertPlaceHolder>" ; } } | org . junit . Assert . assertTrue ( true ) |
shouldBeAbleToRetrieveRestEndpoint ( ) { javax . ws . rs . client . WebTarget target = javax . ws . rs . client . ClientBuilder . newClient ( ) . register ( ws . ament . hammock . johnzon . JohnzonExtension . class ) . target ( ( ( uri ) + "/sse/{uuid}" ) ) . resolveTemplate ( "uuid" , java . util . UUID . randomUUID ( ) . toString ( ) ) ; java . util . List < org . hammock . test . jersey . sse . SseModel > receivedModels = new java . util . ArrayList ( ) ; try ( javax . ws . rs . sse . SseEventSource eventSource = javax . ws . rs . sse . SseEventSource . target ( target ) . build ( ) ) { eventSource . register ( ( event ) -> { org . hammock . test . jersey . sse . SseModel body = event . readData ( . class , MediaType . APPLICATION_JSON_TYPE ) ; System . out . println ( ( "Received<sp>" + ( body . getName ( ) ) ) ) ; receivedModels . add ( body ) ; } , System . out :: println ) ; eventSource . open ( ) ; java . lang . Thread . sleep ( 1000 ) ; } "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertFalse ( receivedModels . isEmpty ( ) ) |
testNotValidateString ( ) { java . lang . String regExStr = "^[\\w]([\\-\\.\\w])+[\\w]+[\\w]*\\.([\\w]+[\\w\\-]+[\\w]*(\\.[a-z][a-z|0-9]*)?)$" ; java . lang . String str = "test.test@gmail.com" ; boolean value = stringUtil . validateString ( regExStr , str ) ; "<AssertPlaceHolder>" ; } validateString ( java . lang . String , java . lang . String ) { try { return regExPatternMatch ( regExStr , str ) ; } catch ( java . lang . Throwable t ) { org . apache . ranger . plugin . model . validation . RangerServiceValidator . LOG . error ( ( ( ( ( ( "Error<sp>validating<sp>string.<sp>str=" + str ) + "<sp>due<sp>to<sp>reason<sp>" ) + ( t . getMessage ( ) ) ) + ".<sp>Stack<sp>Trace<sp>:<sp>" ) + ( t . getStackTrace ( ) ) ) ) ; return false ; } } | org . junit . Assert . assertFalse ( value ) |
withPortStringWhenPortIsEmpty ( ) { final com . azure . common . implementation . http . UrlBuilder builder = new com . azure . common . implementation . http . UrlBuilder ( ) . withPort ( 8080 ) ; builder . withPort ( "" ) ; "<AssertPlaceHolder>" ; } port ( ) { return this . port ; } | org . junit . Assert . assertNull ( builder . port ( ) ) |
Test36 ( ) { cc . iqa . iquery . iQueryParser parser = iQueryIdeParser . createParser ( "><sp>:radio" , false ) ; parser . registerPseudoClass ( "radio" , new cc . iqa . iquery . IPseudoClass ( ) { public boolean resolve ( cc . iqa . iquery . ITreeNode node ) { return iQueryIdeParser . filterByNameEndsWith ( node , "UIASwitch" ) ; } } ) ; java . util . List < cc . iqa . iquery . ITreeNode > candidates = getRoot ( ) ; java . util . List < cc . iqa . iquery . ITreeNode > outTreeNodeList = parser . query ( candidates ) ; "<AssertPlaceHolder>" ; } getRoot ( ) { hierarchyviewerlib . models . LogFileModel model = hierarchyviewerlib . models . LogFileModel . getModel ( ) ; model . loadLogFile ( hierarchyviewerlib . test . AllTests . sLogPath ) ; hierarchyviewerlib . common . ViewNode viewNode = model . getViewNodeByLogName ( hierarchyviewerlib . test . AllTests . sNodeName ) ; java . util . List < cc . iqa . iquery . ITreeNode > candidates = new java . util . ArrayList < cc . iqa . iquery . ITreeNode > ( ) ; candidates . add ( viewNode ) ; return candidates ; } | org . junit . Assert . assertTrue ( ( ( outTreeNodeList . size ( ) ) == 1 ) ) |
fromHostAndPortsWithNullShouldReturnEmptyArray ( ) { org . elasticsearch . common . transport . TransportAddress [ ] result = io . dropwizard . elasticsearch . util . TransportAddressHelper . fromHostAndPorts ( null ) ; "<AssertPlaceHolder>" ; } fromHostAndPorts ( java . util . List ) { if ( hostAndPorts == null ) { return new org . elasticsearch . common . transport . TransportAddress [ 0 ] ; } else { org . elasticsearch . common . transport . TransportAddress [ ] transportAddresses = new org . elasticsearch . common . transport . TransportAddress [ hostAndPorts . size ( ) ] ; for ( int i = 0 ; i < ( hostAndPorts . size ( ) ) ; i ++ ) { transportAddresses [ i ] = io . dropwizard . elasticsearch . util . TransportAddressHelper . fromHostAndPort ( hostAndPorts . get ( i ) ) ; } return transportAddresses ; } } | org . junit . Assert . assertEquals ( 0 , result . length ) |
group_order ( ) { com . querydsl . collections . Map < java . lang . Integer , com . querydsl . core . group . Group > results = com . querydsl . collections . CollQueryFactory . from ( com . querydsl . collections . GroupByTest . post , com . querydsl . collections . GroupByTest . posts ) . from ( com . querydsl . collections . GroupByTest . comment , com . querydsl . collections . GroupByTest . comments ) . where ( com . querydsl . collections . GroupByTest . comment . post . id . eq ( com . querydsl . collections . GroupByTest . post . id ) ) . transform ( groupBy ( com . querydsl . collections . GroupByTest . post . id ) . as ( com . querydsl . collections . GroupByTest . post . name , set ( com . querydsl . collections . GroupByTest . comment . id ) ) ) ; "<AssertPlaceHolder>" ; } size ( ) { query ( store , store . products . size ( ) . gt ( 0 ) ) ; } | org . junit . Assert . assertEquals ( 3 , results . size ( ) ) |
test ( ) { java . lang . String base64EncodeDigest = "aaa" ; eu . europa . esig . dss . DigestDocument doc = new eu . europa . esig . dss . DigestDocument ( ) ; doc . addDigest ( DigestAlgorithm . SHA1 , base64EncodeDigest ) ; "<AssertPlaceHolder>" ; } getDigest ( org . w3c . dom . Element ) { try { org . w3c . dom . NodeList nodeList = eu . europa . esig . dss . DomUtils . getNodeList ( refNode , xPathQueryHolder . XPATH__DS_TRANSFORM ) ; if ( ( nodeList != null ) && ( ( nodeList . getLength ( ) ) > 0 ) ) { throw new eu . europa . esig . dss . DSSException ( "Transformations<sp>are<sp>not<sp>supported" ) ; } java . lang . String digestAlgoUri = eu . europa . esig . dss . DomUtils . getValue ( refNode , xPathQueryHolder . XPATH__DIGEST_METHOD_ALGORITHM ) ; eu . europa . esig . dss . DigestAlgorithm digestAlgorithm = eu . europa . esig . dss . DigestAlgorithm . forXML ( digestAlgoUri ) ; java . lang . String digestValueB64 = eu . europa . esig . dss . DomUtils . getValue ( refNode , xPathQueryHolder . XPATH__DIGEST_VALUE ) ; return new eu . europa . esig . dss . Digest ( digestAlgorithm , eu . europa . esig . dss . utils . Utils . fromBase64 ( digestValueB64 ) ) ; } catch ( java . lang . Exception e ) { eu . europa . esig . dss . xades . validation . ManifestValidator . LOG . warn ( "Unable<sp>to<sp>extract<sp>the<sp>digest<sp>combination<sp>:<sp>{}" , e . getMessage ( ) ) ; return null ; } } | org . junit . Assert . assertEquals ( base64EncodeDigest , doc . getDigest ( DigestAlgorithm . SHA1 ) ) |
capacityViolationAfterAct6_shouldWorkWhenViolated ( ) { buildAnotherScenarioWithOnlyOneVehicleAndWithoutAnyConstraintsBefore ( ) ; jsprit . core . reporting . SolutionPrinter . print ( vrp , solution , SolutionPrinter . Print . VERBOSE ) ; jsprit . core . analysis . SolutionAnalyser analyser = new jsprit . core . analysis . SolutionAnalyser ( vrp , solution , new jsprit . core . problem . cost . TransportDistance ( ) { @ jsprit . core . analysis . Override public double getDistance ( jsprit . core . problem . Location from , jsprit . core . problem . Location to ) { return vrp . getTransportCosts ( ) . getTransportCost ( from , to , 0.0 , null , null ) ; } } ) ; jsprit . core . problem . solution . route . VehicleRoute route = solution . getRoutes ( ) . iterator ( ) . next ( ) ; jsprit . core . problem . Capacity cap = analyser . getCapacityViolationAfterActivity ( route . getActivities ( ) . get ( 5 ) , route ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( 10 , cap . get ( 0 ) ) |
testCanProceedWhenThresholdLessThanConfigured ( ) { impl . setRequestPerSecondThreshold ( 50 ) ; "<AssertPlaceHolder>" ; } canProceed ( ) { if ( ( requestPerSecondThreshold ) == ( - 1 ) ) { return true ; } if ( ( lastRequestOccurance ) == 0 ) { lastRequestOccurance = java . lang . System . currentTimeMillis ( ) ; return true ; } if ( ( ( java . lang . System . currentTimeMillis ( ) ) - ( lastRequestOccurance ) ) > ( deltaWaitTimeMillis ) ) { lastRequestOccurance = java . lang . System . currentTimeMillis ( ) ; return true ; } return false ; } | org . junit . Assert . assertTrue ( impl . canProceed ( ) ) |
testName ( ) { font . setName ( "huhu" ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertEquals ( "huhu" , font . getName ( ) ) |
testCompareLayerMethods ( ) { org . deeplearning4j . nn . multilayer . MultiLayerConfiguration conf1 = new org . deeplearning4j . nn . multilayer . NeuralNetConfiguration . Builder ( ) . seed ( 123 ) . list ( ) . layer ( 0 , new org . deeplearning4j . nn . multilayer . DenseLayer . Builder ( ) . nIn ( 4 ) . nOut ( 3 ) . weightInit ( WeightInit . XAVIER ) . activation ( Activation . TANH ) . build ( ) ) . layer ( 1 , new org . deeplearning4j . nn . multilayer . DenseLayer . Builder ( ) . nIn ( 3 ) . nOut ( 2 ) . weightInit ( WeightInit . XAVIER ) . activation ( Activation . TANH ) . build ( ) ) . layer ( 2 , new org . deeplearning4j . nn . multilayer . LSTM . Builder ( ) . nIn ( 2 ) . nOut ( 2 ) . build ( ) ) . layer ( 3 , new org . deeplearning4j . nn . multilayer . RnnOutputLayer . Builder ( LossFunctions . LossFunction . MCXENT ) . weightInit ( WeightInit . XAVIER ) . activation ( Activation . SOFTMAX ) . nIn ( 2 ) . nOut ( 3 ) . build ( ) ) . build ( ) ; org . deeplearning4j . nn . multilayer . MultiLayerConfiguration conf2 = new org . deeplearning4j . nn . multilayer . NeuralNetConfiguration . Builder ( ) . seed ( 123 ) . list ( ) . layer ( new org . deeplearning4j . nn . multilayer . DenseLayer . Builder ( ) . nIn ( 4 ) . nOut ( 3 ) . weightInit ( WeightInit . XAVIER ) . activation ( Activation . TANH ) . build ( ) ) . layer ( new org . deeplearning4j . nn . multilayer . DenseLayer . Builder ( ) . nIn ( 3 ) . nOut ( 2 ) . weightInit ( WeightInit . XAVIER ) . activation ( Activation . TANH ) . build ( ) ) . layer ( new org . deeplearning4j . nn . multilayer . LSTM . Builder ( ) . nIn ( 2 ) . nOut ( 2 ) . build ( ) ) . layer ( new org . deeplearning4j . nn . multilayer . RnnOutputLayer . Builder ( LossFunctions . LossFunction . MCXENT ) . weightInit ( WeightInit . XAVIER ) . activation ( Activation . SOFTMAX ) . nIn ( 2 ) . nOut ( 3 ) . build ( ) ) . build ( ) ; "<AssertPlaceHolder>" ; } build ( ) { return new org . datavec . api . records . reader . impl . jackson . FieldSelection ( this ) ; } | org . junit . Assert . assertEquals ( conf1 , conf2 ) |
whenString ( ) { allowingObjectSpecCorrespondingClassIs ( java . lang . String . class ) ; allowingObjectSpecHas ( org . apache . isis . core . metamodel . facets . object . encodeable . EncodableFacet . class , mockEncodableFacet ) ; context . checking ( new org . jmock . Expectations ( ) { { oneOf ( mockObjectAdapter ) . getObject ( ) ; will ( returnValue ( "encodedString" ) ) ; } } ) ; final java . lang . Object actual = org . apache . isis . viewer . restfulobjects . rendering . domainobjects . JsonValueEncoder . asObject ( mockObjectAdapter , null ) ; "<AssertPlaceHolder>" ; } getObject ( ) { return pojo ; } | org . junit . Assert . assertSame ( "encodedString" , actual ) |
loadCatchesUnforeseenExceptions ( ) { java . lang . Exception testException = new java . lang . RuntimeException ( "Test<sp>Exception" ) ; doThrow ( testException ) . when ( this . doc ) . getXObject ( Patient . CLASS_REFERENCE ) ; org . phenotips . data . PatientData < java . lang . String > result = this . component . load ( this . patient ) ; verify ( this . mocker . getMockedLogger ( ) ) . error ( eq ( PatientDataController . ERROR_MESSAGE_LOAD_FAILED ) , anyString ( ) ) ; "<AssertPlaceHolder>" ; } load ( org . xwiki . bridge . DocumentModelBridge ) { try { return getEntityConstructor ( ) . newInstance ( document ) ; } catch ( java . lang . IllegalArgumentException | java . lang . reflect . InvocationTargetException ex ) { this . logger . info ( "Tried<sp>to<sp>load<sp>invalid<sp>entity<sp>of<sp>type<sp>[{}]<sp>from<sp>document<sp>[{}]" , getEntityXClassReference ( ) , ( document == null ? null : document . getDocumentReference ( ) ) ) ; } catch ( java . lang . InstantiationException | java . lang . IllegalAccessException ex ) { this . logger . error ( "Failed<sp>to<sp>instantiate<sp>primary<sp>entity<sp>of<sp>type<sp>[{}]<sp>from<sp>document<sp>[{}]:<sp>{}" , getEntityXClassReference ( ) , ( document == null ? null : document . getDocumentReference ( ) ) , ex . getMessage ( ) ) ; } return null ; } | org . junit . Assert . assertNull ( result ) |
testAdler ( ) { java . util . Random rnd = new java . util . Random ( 42 ) ; for ( int i = 0 ; i < 100 ; i ++ ) { byte [ ] test = new byte [ 1000 + i ] ; rnd . nextBytes ( test ) ; int start = rnd . nextInt ( test . length ) ; int len = rnd . nextInt ( ( ( test . length ) - start ) ) ; net . tomp2p . synchronization . RSync . RollingChecksum a = new net . tomp2p . synchronization . RSync . RollingChecksum ( ) ; a . update ( test , start , len ) ; int val = a . value ( ) ; a . updateRolling ( test ) ; val = a . value ( ) ; a . reset ( ) ; a . update ( test , ( start + 1 ) , len ) ; int val2 = a . value ( ) ; "<AssertPlaceHolder>" ; } } value ( ) { return ( ( b ) << 16 ) | ( a ) ; } | org . junit . Assert . assertEquals ( val , val2 ) |
testParsingIpv4Address ( ) { uri = "postgresql://192.168.1.1" ; actualParsedResult = parse ( uri ) ; expectedParsedResult = new io . vertx . core . json . JsonObject ( ) . put ( "host" , "192.168.1.1" ) ; "<AssertPlaceHolder>" ; } parse ( java . lang . String$String ) { try { io . vertx . core . json . JsonObject configuration = new io . vertx . core . json . JsonObject ( ) ; io . reactiverse . pgclient . impl . PgConnectionUriParser . doParse ( connectionUri , configuration ) ; return configuration ; } catch ( java . lang . Exception e ) { throw new java . lang . IllegalArgumentException ( ( "Cannot<sp>parse<sp>invalid<sp>connection<sp>URI:<sp>" + connectionUri ) , e ) ; } } | org . junit . Assert . assertEquals ( expectedParsedResult , actualParsedResult ) |
testDynamicRuleRemovalsSubNetworkAndNot ( ) { org . drools . core . impl . InternalKnowledgeBase kbase = ( ( org . drools . core . impl . InternalKnowledgeBase ) ( loadKnowledgeBase ( "test_DynamicRulesWithNotSubnetwork.drl" ) ) ) ; org . kie . api . runtime . KieSession ksession = createKnowledgeSession ( kbase ) ; final org . kie . api . event . rule . AgendaEventListener alistener = mock ( org . kie . api . event . rule . AgendaEventListener . class ) ; ksession . addEventListener ( alistener ) ; ksession . insert ( new org . drools . compiler . Person ( "toni" ) ) ; ksession . fireAllRules ( ) ; verify ( alistener , never ( ) ) . matchCreated ( any ( org . kie . api . event . rule . MatchCreatedEvent . class ) ) ; ksession . insert ( new org . drools . compiler . Person ( "bob" ) ) ; ksession . fireAllRules ( ) ; verify ( alistener , times ( 1 ) ) . matchCreated ( any ( org . kie . api . event . rule . MatchCreatedEvent . class ) ) ; ksession . insert ( new org . drools . compiler . Person ( "mark" ) ) ; ksession . fireAllRules ( ) ; verify ( alistener , times ( 1 ) ) . matchCreated ( any ( org . kie . api . event . rule . MatchCreatedEvent . class ) ) ; kbase . removeKiePackage ( "org.drools.compiler" ) ; "<AssertPlaceHolder>" ; java . util . Collection < org . kie . api . definition . KiePackage > kpkgs = loadKnowledgePackages ( "test_DynamicRulesWithNotSubnetwork.drl" ) ; kbase . addPackages ( kpkgs ) ; ksession . fireAllRules ( ) ; verify ( alistener , times ( 2 ) ) . matchCreated ( any ( org . kie . api . event . rule . MatchCreatedEvent . class ) ) ; } getKiePackages ( ) { org . drools . core . definitions . InternalKnowledgePackage [ ] pkgs = getPackages ( ) ; java . util . List < org . kie . api . definition . KiePackage > list = new java . util . ArrayList < org . kie . api . definition . KiePackage > ( pkgs . length ) ; java . util . Collections . addAll ( list , pkgs ) ; return list ; } | org . junit . Assert . assertEquals ( 0 , kbase . getKiePackages ( ) . size ( ) ) |
testUploadCancellation ( ) { java . lang . String user = "user1" ; org . genemania . engine . cache . RandomDataCacheBuilder rcb = new org . genemania . engine . cache . RandomDataCacheBuilder ( 2112 ) ; rcb . setUp ( ) ; long [ ] organism1NetworkIds = rcb . addOrganism ( 1 , 20 , 5 , 0.5 , 5 , 0.5 ) ; long [ ] organism1NodeIds = rcb . getCache ( ) . getNodeIds ( 1 ) . getNodeIds ( ) ; org . genemania . dto . UploadNetworkEngineRequestDto request = new org . genemania . dto . UploadNetworkEngineRequestDto ( ) ; request . setLayout ( DataLayout . WEIGHTED_NETWORK ) ; request . setMethod ( NetworkProcessingMethod . DIRECT ) ; request . setNamespace ( user ) ; request . setProgressReporter ( org . genemania . util . NullProgressReporter . instance ( ) ) ; request . setOrganismId ( 1 ) ; request . setNetworkId ( ( - 2 ) ) ; request . setSparsification ( 50 ) ; java . lang . String data = "10000\t10001\t1.0\n10000\t10002\t0.8\n" ; request . setData ( new java . io . StringReader ( data ) ) ; org . genemania . util . ProgressReporter progress = new org . genemania . engine . SimpleProgressReporter ( ) ; progress . cancel ( ) ; request . setProgressReporter ( progress ) ; org . genemania . engine . Mania2 mania = new org . genemania . engine . Mania2 ( rcb . getCache ( ) ) ; org . genemania . dto . UploadNetworkEngineResponseDto result = mania . uploadNetwork ( request ) ; "<AssertPlaceHolder>" ; } uploadNetwork ( org . genemania . dto . UploadNetworkWebRequestDto ) { org . genemania . broker . SyncWorker . UploadResult ret = new org . genemania . broker . SyncWorker . UploadResult ( ) ; org . genemania . dto . UploadNetworkEngineRequestDto engReq = ret . request = org . genemania . util . BrokerUtils . dto2dto ( req ) ; org . genemania . dto . UploadNetworkEngineResponseDto engRes = ret . response = engine . uploadNetwork ( engReq ) ; ret . request = engReq ; ret . response = engRes ; return ret ; } | org . junit . Assert . assertNull ( result ) |
convergeCreatesIfAbsent ( ) { intent . setResolution ( com . engineerbetter . converger . resolution . IdentifiableResolution . absent ( ) ) ; given ( uaa . createUser ( intent . properties ) ) . willReturn ( "new-id" ) ; handler . converge ( ) ; verify ( uaa ) . createUser ( new com . engineerbetter . converger . properties . UaaUserProperties ( "test@example.com" , "Geoff" , "McGeofferson" ) ) ; "<AssertPlaceHolder>" ; } getResolution ( ) { return resolution ; } | org . junit . Assert . assertThat ( intent . getResolution ( ) . exists ( ) , is ( true ) ) |
testGetAttribute ( ) { org . hisp . dhis . attribute . Attribute attribute = new org . hisp . dhis . attribute . Attribute ( ) ; attribute . setValueType ( ValueType . TEXT ) ; attribute . setName ( "attribute1" ) ; attributeService . addAttribute ( attribute ) ; attribute = attributeService . getAttribute ( attribute . getId ( ) ) ; "<AssertPlaceHolder>" ; } getId ( ) { return id ; } | org . junit . Assert . assertNotNull ( attribute ) |
toFromMapTest ( ) { datawave . webservice . common . audit . AuditParameters auditParams = new datawave . webservice . common . audit . AuditParameters ( ) ; auditParams . setUserDn ( "someUser" ) ; auditParams . setQuery ( "someQuery" ) ; auditParams . setSelectors ( datawave . webservice . common . audit . Arrays . asList ( "sel1" , "sel2" ) ) ; auditParams . setAuths ( "AUTH1,AUTH2" ) ; auditParams . setAuditType ( Auditor . AuditType . ACTIVE ) ; auditParams . setColviz ( new org . apache . accumulo . core . security . ColumnVisibility ( "ALL" ) ) ; auditParams . setQueryDate ( new datawave . webservice . common . audit . Date ( ) ) ; auditParams . setAuditId ( datawave . webservice . common . audit . UUID . randomUUID ( ) . toString ( ) ) ; datawave . webservice . common . audit . AuditParameters fromMapParams = auditParams . fromMap ( auditParams . toMap ( ) ) ; "<AssertPlaceHolder>" ; } toMap ( ) { return java . util . Collections . singletonMap ( datawave . marking . ColumnVisibilitySecurityMarking . VISIBILITY_MARKING , this . columnVisibility ) ; } | org . junit . Assert . assertEquals ( auditParams , fromMapParams ) |
testIsTransformableWithNonBlockSizeMultipleDimensions ( ) { "<AssertPlaceHolder>" ; } isTransformable ( ) { return ( ( ( ( getWidth ( ) ) >= ( edu . illinois . library . cantaloupe . processor . codec . jpeg . TurboJPEGImageReader . TRANSFORM_MIN_SIZE ) ) && ( ( getHeight ( ) ) >= ( edu . illinois . library . cantaloupe . processor . codec . jpeg . TurboJPEGImageReader . TRANSFORM_MIN_SIZE ) ) ) && ( ( ( getWidth ( ) ) % ( getBlockWidth ( ) ) ) == 0 ) ) && ( ( ( getHeight ( ) ) % ( getBlockHeight ( ) ) ) == 0 ) ; } | org . junit . Assert . assertFalse ( instance . isTransformable ( ) ) |
foldRight_A$Object$Function2_Nil ( ) { com . m3 . scalaflavor4j . Seq < java . lang . Integer > seq = com . m3 . scalaflavor4j . Seq . apply ( ) ; java . lang . String actual = seq . foldRight ( "foo" , new com . m3 . scalaflavor4j . F2 < java . lang . Integer , java . lang . String , java . lang . String > ( ) { public java . lang . String apply ( java . lang . Integer v1 , java . lang . String v2 ) { return v1 + v2 ; } } ) ; "<AssertPlaceHolder>" ; } apply ( ) { com . m3 . scalaflavor4j . MainFunction main = new com . m3 . scalaflavor4j . MainFunction ( ) { public void apply ( java . lang . String [ ] args ) throws com . m3 . scalaflavor4j . Exception { print . apply ( args . length ) ; com . m3 . scalaflavor4j . Seq . apply ( args ) . foreach ( new com . m3 . scalaflavor4j . VoidF1 < java . lang . String > ( ) { public void apply ( java . lang . String arg ) throws com . m3 . scalaflavor4j . Exception { print . apply ( arg ) ; } } ) ; } } ; main . apply ( new java . lang . String [ ] { "a" , "b" } ) ; } | org . junit . Assert . assertThat ( actual , org . hamcrest . CoreMatchers . is ( "foo" ) ) |
testIncrementEvaluations5 ( ) { instance . setup ( instance , null ) ; instance . evaluations = 0 ; instance . evalthreads = 5 ; instance . incrementEvaluations ( 1 ) ; "<AssertPlaceHolder>" ; } incrementEvaluations ( int ) { if ( ( evalthreads ) == 1 ) evaluations += val ; else { synchronizedIncrementEvaluations ( val ) ; } } | org . junit . Assert . assertEquals ( 1 , instance . evaluations ) |
testGetValueCookiesDate ( ) { java . util . Date date = new java . util . Date ( ) ; java . lang . String strDate = com . fasterxml . jackson . databind . util . ISO8601Utils . format ( date ) ; javax . servlet . http . Cookie [ ] cookies = new javax . servlet . http . Cookie [ ] { new javax . servlet . http . Cookie ( "c1" , strDate ) } ; new mockit . Expectations ( ) { { request . getCookies ( ) ; result = cookies ; } } ; org . apache . servicecomb . common . rest . codec . param . CookieProcessorCreator . CookieProcessor processor = createProcessor ( "c1" , java . util . Date . class ) ; java . lang . Object value = processor . getValue ( request ) ; "<AssertPlaceHolder>" ; } getValue ( java . lang . Object [ ] ) { return ( ( T ) ( args [ paramIndex ] ) ) ; } | org . junit . Assert . assertEquals ( strDate , com . fasterxml . jackson . databind . util . ISO8601Utils . format ( ( ( java . util . Date ) ( value ) ) ) ) |
testGetBorder ( ) { try { org . odftoolkit . simple . style . Border borderbase = new org . odftoolkit . simple . style . Border ( org . odftoolkit . odfdom . type . Color . LIME , 3.0701 , 0.0208 , 0.0346 , org . odftoolkit . simple . style . StyleTypeDefinitions . SupportedLinearMeasure . CM ) ; borderbase . setLineStyle ( StyleTypeDefinitions . LineType . SINGLE ) ; org . odftoolkit . simple . SpreadsheetDocument doc = org . odftoolkit . simple . SpreadsheetDocument . newSpreadsheetDocument ( ) ; org . odftoolkit . simple . table . Table table = doc . getTableByName ( "Sheet1" ) ; org . odftoolkit . simple . table . Cell cell = table . getCellByPosition ( 2 , 2 ) ; cell . setBorders ( CellBordersType . ALL_FOUR , borderbase ) ; org . odftoolkit . simple . style . TableCellProperties styleCell = cell . getStyleHandler ( ) . getTableCellPropertiesForWrite ( ) ; org . odftoolkit . simple . style . Border bor = styleCell . getBorder ( ) ; "<AssertPlaceHolder>" ; doc . save ( org . odftoolkit . simple . utils . ResourceUtilities . newTestOutputFile ( "testSupportedLinearMeasure.ods" ) ) ; } catch ( java . lang . Exception e ) { org . odftoolkit . simple . style . TableCellPropertiesTest . LOGGER . log ( Level . SEVERE , e . getMessage ( ) , e ) ; org . junit . Assert . fail ( e . getMessage ( ) ) ; } } getBorder ( ) { return mBorderPropertiesHandler . getBorder ( ) ; } | org . junit . Assert . assertEquals ( borderbase , bor ) |
deletionOfAnUnexistingResource ( ) { try { deleteAt ( anUnexistingResourceURI ( ) ) ; } catch ( javax . ws . rs . WebApplicationException ex ) { int receivedStatus = ex . getResponse ( ) . getStatus ( ) ; int notFound = Status . NOT_FOUND . getStatusCode ( ) ; "<AssertPlaceHolder>" ; } } is ( T ) { return java . util . Objects . equals ( this . value , value ) ; } | org . junit . Assert . assertThat ( receivedStatus , org . hamcrest . Matchers . is ( notFound ) ) |
testCanAddNewDependencyOfLowerType ( ) { final org . springframework . roo . project . Dependency mockNewDependency = mock ( org . springframework . roo . project . Dependency . class ) ; when ( mockNewDependency . getType ( ) ) . thenReturn ( DependencyType . JAR ) ; final org . springframework . roo . project . maven . Pom pom = getMinimalPom ( org . springframework . roo . project . maven . PomTest . WAR ) ; "<AssertPlaceHolder>" ; } canAddDependency ( org . springframework . roo . project . Dependency , boolean ) { return ( ( newDependency != null ) && ( ! ( isDependencyRegistered ( newDependency , checkVersion ) ) ) ) && ( ! ( org . springframework . roo . project . Dependency . isHigherLevel ( newDependency . getType ( ) . toString ( ) , packaging ) ) ) ; } | org . junit . Assert . assertTrue ( pom . canAddDependency ( mockNewDependency , true ) ) |
testMergeFieldsAfterFieldValueIsRemovedFromTheMiddleOfSeries ( ) { com . liferay . dynamic . data . mapping . model . DDMForm ddmForm = createDDMForm ( ) ; addDDMFormFields ( ddmForm , createTextDDMFormField ( "Content" , "" , true , true , false ) ) ; com . liferay . dynamic . data . mapping . model . DDMStructure ddmStructure = createStructure ( "Test<sp>Structure" , ddmForm ) ; com . liferay . dynamic . data . mapping . storage . Field existingContentField = createField ( ddmStructure . getStructureId ( ) , "Content" , createValuesList ( "Content<sp>1" , "Content<sp>1" 0 , "Content<sp>3" ) , createValuesList ( "Conteudo<sp>1" , "Conteudo<sp>2<sp>" , "Conteudo<sp>3" ) ) ; com . liferay . dynamic . data . mapping . storage . Field existingFieldsDisplayField = createFieldsDisplayField ( ddmStructure . getStructureId ( ) , ( "Content_INSTANCE_ovho,Content_INSTANCE_zuvh," + "Content_INSTANCE_yiek" ) ) ; com . liferay . dynamic . data . mapping . storage . Fields existingFields = createFields ( existingContentField , existingFieldsDisplayField ) ; com . liferay . dynamic . data . mapping . storage . Field newContentField = createField ( ddmStructure . getStructureId ( ) , "Content" , createValuesList ( "Content<sp>1" , "Content<sp>3" ) , null ) ; com . liferay . dynamic . data . mapping . storage . Field newFieldsDisplayField = createFieldsDisplayField ( ddmStructure . getStructureId ( ) , "Content<sp>1" 1 ) ; com . liferay . dynamic . data . mapping . storage . Fields newFields = createFields ( newContentField , newFieldsDisplayField ) ; com . liferay . dynamic . data . mapping . storage . Fields actualFields = _ddm . mergeFields ( newFields , existingFields ) ; com . liferay . dynamic . data . mapping . storage . Field actualContentField = actualFields . get ( "Content" ) ; "<AssertPlaceHolder>" ; testValues ( actualContentField . getValues ( LocaleUtil . US ) , "Content<sp>1" , "Content<sp>3" ) ; testValues ( actualContentField . getValues ( LocaleUtil . BRAZIL ) , "Conteudo<sp>1" , "Conteudo<sp>3" ) ; } get ( java . lang . String ) { return com . liferay . journal . content . search . web . configuration . JournalContentSearchWebConfigurationUtil . _configuration . get ( key ) ; } | org . junit . Assert . assertNotNull ( actualContentField ) |
bitfieldShouldAllowMultipleSubcommands ( ) { actual . add ( connection . bitfield ( org . springframework . data . redis . connection . KEY_1 , create ( ) . incr ( signed ( 5 ) ) . valueAt ( offset ( 100L ) ) . by ( 1L ) . get ( unsigned ( 4 ) ) . valueAt ( 0L ) ) ) ; "<AssertPlaceHolder>" ; } getResults ( ) { connection . exec ( ) ; java . util . List < java . lang . Object > txResults = ( ( java . util . List < java . lang . Object > ) ( connection . closePipeline ( ) . get ( 0 ) ) ) ; return txResults ; } | org . junit . Assert . assertThat ( ( ( org . springframework . data . redis . connection . List < java . lang . Long > ) ( getResults ( ) . get ( 0 ) ) ) , contains ( 1L , 0L ) ) |
testValidationWithOldConfiguration ( ) { java . nio . file . Path schemaPath = java . nio . file . Paths . get ( java . lang . Thread . currentThread ( ) . getContextClassLoader ( ) . getResource ( "generation/simpleSchema.json" ) . toURI ( ) ) ; org . kaaproject . kaa . server . common . core . schema . BaseSchema configuraionSchema = new org . kaaproject . kaa . server . common . core . schema . BaseSchema ( new java . lang . String ( java . nio . file . Files . readAllBytes ( schemaPath ) ) ) ; org . apache . avro . Schema avroSchema = new org . apache . avro . Schema . Parser ( ) . parse ( configuraionSchema . getRawSchema ( ) ) ; org . apache . avro . generic . GenericRecord recordNew = new org . apache . avro . generic . GenericData . Record ( avroSchema ) ; recordNew . put ( "intField" , 4 ) ; org . apache . avro . generic . GenericFixed uuidNew = org . kaaproject . kaa . server . common . core . algorithms . AvroUtils . generateUuidObject ( ) ; recordNew . put ( CommonConstants . UUID_FIELD , uuidNew ) ; org . apache . avro . generic . GenericRecord recordOld = new org . apache . avro . generic . GenericData . Record ( avroSchema ) ; recordOld . put ( "intField" , 5 ) ; org . apache . avro . generic . GenericFixed uuidOld = org . kaaproject . kaa . server . common . core . algorithms . AvroUtils . generateUuidObject ( ) ; recordOld . put ( CommonConstants . UUID_FIELD , uuidOld ) ; org . kaaproject . kaa . common . avro . GenericAvroConverter < org . apache . avro . generic . GenericRecord > converter = new org . kaaproject . kaa . common . avro . GenericAvroConverter ( avroSchema ) ; java . lang . String configurationBodyNew = converter . encodeToJson ( recordNew ) ; java . lang . String configurationBodyOld = converter . encodeToJson ( recordOld ) ; org . kaaproject . kaa . server . common . core . algorithms . validator . DefaultUuidValidator uuidGenerator = new org . kaaproject . kaa . server . common . core . algorithms . validator . DefaultUuidValidator ( configuraionSchema , new org . kaaproject . kaa . server . common . core . configuration . BaseDataFactory ( ) ) ; org . kaaproject . kaa . server . common . core . configuration . KaaData processedConfigurationBody = uuidGenerator . validateUuidFields ( recordNew , recordOld ) ; org . apache . avro . generic . GenericRecord processedConfiguration = converter . decodeJson ( processedConfigurationBody . getRawData ( ) ) ; "<AssertPlaceHolder>" ; } get ( int ) { switch ( field$ ) { case 0 : return extensionId ; case 1 : return payload ; default : throw new org . apache . avro . AvroRuntimeException ( "Bad<sp>index" ) ; } } | org . junit . Assert . assertEquals ( processedConfiguration . get ( CommonConstants . UUID_FIELD ) , uuidOld ) |
testPartialCopyInputStreamOutputStream ( ) { try ( java . io . InputStream bufferedSource = new java . io . ByteArrayInputStream ( sourceBytes ) ; java . io . InputStream source = new net . yacy . kelondro . util . FileUtilsTest . LowerReadThanExpectedInputStream ( bufferedSource ) ; java . io . ByteArrayOutputStream dest = new java . io . ByteArrayOutputStream ( ) ) { net . yacy . kelondro . util . FileUtils . copy ( source , dest , COUNT ) ; "<AssertPlaceHolder>" ; } } size ( ) { return elt ; } | org . junit . Assert . assertEquals ( COUNT , dest . size ( ) ) |
privateConstructor ( ) { final org . apache . johnzon . mapper . MapperTest . HiddenConstructor value = new org . apache . johnzon . mapper . MapperBuilder ( ) . setSupportHiddenAccess ( true ) . build ( ) . readObject ( new java . io . ByteArrayInputStream ( "{\"value\":1}" . getBytes ( ) ) , org . apache . johnzon . mapper . MapperTest . HiddenConstructor . class ) ; "<AssertPlaceHolder>" ; } readObject ( javax . json . JsonStructure , java . lang . reflect . Type ) { return new org . apache . johnzon . mapper . MappingParserImpl ( config , mappings , new javax . json . JsonReader ( ) { @ org . apache . johnzon . mapper . Override public javax . json . JsonStructure read ( ) { return value ; } @ org . apache . johnzon . mapper . Override public javax . json . JsonValue readValue ( ) { return value ; } @ org . apache . johnzon . mapper . Override public javax . json . JsonObject readObject ( ) { return value . asJsonObject ( ) ; } @ org . apache . johnzon . mapper . Override public javax . json . JsonArray readArray ( ) { return value . asJsonArray ( ) ; } @ org . apache . johnzon . mapper . Override public void close ( ) { } } , isDedup ( clazz ) ) . readObject ( clazz ) ; } | org . junit . Assert . assertEquals ( 1 , value . value ) |
datumEindeDeelsOnbekend ( ) { final nl . bzk . algemeenbrp . dal . domein . brp . entity . BijhouderFiatteringsuitzondering bijhouderFiatteringsuitzondering = new nl . bzk . algemeenbrp . dal . domein . brp . entity . BijhouderFiatteringsuitzondering ( new nl . bzk . algemeenbrp . dal . domein . brp . entity . PartijRol ( new nl . bzk . algemeenbrp . dal . domein . brp . entity . Partij ( "partij1" , "000001" ) , nl . bzk . algemeenbrp . dal . domein . brp . enums . Rol . AFNEMER ) ) ; bijhouderFiatteringsuitzondering . setDatumIngang ( 20160005 ) ; "<AssertPlaceHolder>" ; } errorCount ( nl . bzk . algemeenbrp . dal . domein . brp . entity . ToegangLeveringsAutorisatie ) { final java . util . Map < ? , ? > result = new java . util . HashMap ( ) ; final org . springframework . validation . Errors errors = new org . springframework . validation . MapBindingResult ( result , "toegang" ) ; subject . validate ( toegang , errors ) ; return errors . getErrorCount ( ) ; } | org . junit . Assert . assertEquals ( 1 , errorCount ( bijhouderFiatteringsuitzondering ) ) |
canBeSerializedAndDeserialized ( ) { com . eclipsesource . json . JsonString string = new com . eclipsesource . json . JsonString ( "foo" ) ; "<AssertPlaceHolder>" ; } serializeAndDeserialize ( T ) { return ( ( T ) ( com . eclipsesource . json . TestUtil . deserialize ( com . eclipsesource . json . TestUtil . serialize ( instance ) ) ) ) ; } | org . junit . Assert . assertEquals ( string , com . eclipsesource . json . TestUtil . serializeAndDeserialize ( string ) ) |
testTableMapReducer ( ) { final org . apache . hadoop . conf . Configuration conf = createConfiguration ( ) ; final org . apache . hadoop . fs . FileSystem fs = org . apache . hadoop . fs . FileSystem . get ( conf ) ; final org . kiji . schema . KijiURI uri = getKijiURI ( ) ; final org . kiji . schema . Kiji kiji = Kiji . Factory . open ( uri , conf ) ; try { final int nregions = 16 ; final org . kiji . schema . avro . TableLayoutDesc layout = org . kiji . mapreduce . KijiMRTestLayouts . getTestLayout ( ) ; final java . lang . String tableName = layout . getName ( ) ; kiji . createTable ( layout , nregions ) ; final org . kiji . schema . KijiTable table = kiji . openTable ( tableName ) ; try { final org . kiji . schema . KijiTableWriter writer = table . openTableWriter ( ) ; try { for ( int i = 0 ; i < 10 ; ++ i ) { writer . put ( table . getEntityId ( ( "row-" + i ) ) , "primitives" , "int" , ( i % 3 ) ) ; } } finally { writer . close ( ) ; } final org . apache . hadoop . fs . Path output = new org . apache . hadoop . fs . Path ( fs . getUri ( ) . toString ( ) , java . lang . String . format ( "/%s-%s-%d/table-mr-output" , getClass ( ) . getName ( ) , mTestName . getMethodName ( ) , java . lang . System . currentTimeMillis ( ) ) ) ; final org . kiji . mapreduce . KijiMapReduceJob mrjob = org . kiji . mapreduce . gather . KijiGatherJobBuilder . create ( ) . withConf ( conf ) . withGatherer ( SimpleTableMapReducer . TableMapper . class ) . withReducer ( SimpleTableMapReducer . TableReducer . class ) . withInputTable ( table . getURI ( ) ) . withOutput ( org . kiji . mapreduce . output . MapReduceJobOutputs . newHFileMapReduceJobOutput ( table . getURI ( ) , output , 16 ) ) . build ( ) ; "<AssertPlaceHolder>" ; } finally { table . release ( ) ; } } finally { kiji . release ( ) ; } } run ( ) { try { mapper . run ( subcontext ) ; } catch ( java . lang . Throwable ie ) { throwable = ie ; } } | org . junit . Assert . assertTrue ( mrjob . run ( ) ) |
testPrint ( ) { aima . core . environment . nqueens . NQueensBoard board2 = new aima . core . environment . nqueens . NQueensBoard ( 2 ) ; board2 . addQueenAt ( new aima . core . util . datastructure . XYLocation ( 0 , 0 ) ) ; java . lang . String expected = "<sp>Q<sp>-<sp>\n<sp>-<sp>-<sp>\n" ; "<AssertPlaceHolder>" ; } getBoardPic ( ) { java . lang . StringBuilder builder = new java . lang . StringBuilder ( ) ; for ( int row = 0 ; row < ( getSize ( ) ) ; row ++ ) { for ( int col = 0 ; col < ( getSize ( ) ) ; col ++ ) { if ( queenExistsAt ( col , row ) ) builder . append ( "<sp>Q<sp>" ) ; else builder . append ( "<sp>-<sp>" ) ; } builder . append ( "\n" ) ; } return builder . toString ( ) ; } | org . junit . Assert . assertEquals ( expected , board2 . getBoardPic ( ) ) |
testImageUri ( ) { org . apache . olingo . odata2 . ref . model . Employee employee1 = new org . apache . olingo . odata2 . ref . model . Employee ( 1 , null ) ; employee1 . setImageUri ( org . apache . olingo . odata2 . ref . model . EmployeeTest . VALUE_URI ) ; "<AssertPlaceHolder>" ; } getImageUri ( ) { return imageUrl ; } | org . junit . Assert . assertEquals ( org . apache . olingo . odata2 . ref . model . EmployeeTest . VALUE_URI , employee1 . getImageUri ( ) ) |
testDefaultReadWriteComposite ( ) { init ( null ) ; javax . xml . stream . XMLStreamReader reader = inputFactory . createXMLStreamReader ( new java . io . StringReader ( org . apache . tuscany . sca . assembly . xml . ReadWriteBindingSCATestCase . XML ) ) ; org . apache . tuscany . sca . assembly . Composite composite = ( ( org . apache . tuscany . sca . assembly . Composite ) ( staxProcessor . read ( reader , context ) ) ) ; "<AssertPlaceHolder>" ; reader . close ( ) ; verifyComposite ( composite ) ; java . io . ByteArrayOutputStream bos = new java . io . ByteArrayOutputStream ( ) ; staxProcessor . write ( composite , bos , context ) ; java . io . ByteArrayInputStream bis = new java . io . ByteArrayInputStream ( bos . toByteArray ( ) ) ; composite = staxProcessor . read ( bis , org . apache . tuscany . sca . assembly . Composite . class , context ) ; verifyComposite ( composite ) ; } read ( javax . xml . stream . XMLStreamReader , org . apache . tuscany . sca . contribution . processor . ProcessorContext ) { org . apache . tuscany . sca . binding . ws . axis2 . policy . configuration . Axis2ConfigParamPolicy policy = new org . apache . tuscany . sca . binding . ws . axis2 . policy . configuration . Axis2ConfigParamPolicy ( ) ; int event = reader . getEventType ( ) ; javax . xml . namespace . QName name = null ; org . apache . axiom . om . OMElement parameterElement = null ; java . lang . String paramName = null ; while ( reader . hasNext ( ) ) { event = reader . getEventType ( ) ; switch ( event ) { case START_ELEMENT : { name = reader . getName ( ) ; if ( org . apache . tuscany . sca . binding . ws . axis2 . policy . configuration . Axis2ConfigParamPolicyProcessor . PARAMETER . equals ( name . getLocalPart ( ) ) ) { paramName = reader . getAttributeValue ( null , "name" ) ; parameterElement = loadElement ( reader ) ; policy . getParamElements ( ) . put ( paramName , parameterElement ) ; } break ; } } if ( event == ( END_ELEMENT ) ) { if ( org . apache . tuscany . sca . binding . ws . axis2 . policy . configuration . Axis2ConfigParamPolicyProcessor . AXIS2_CONFIG_PARAM_POLICY_QNAME . equals ( reader . getName ( ) ) ) { break ; } } if ( reader . hasNext ( ) ) { reader . next ( ) ; } } return policy ; } | org . junit . Assert . assertNotNull ( composite ) |
shouldThrowIfTransactionFailed ( ) { final com . spotify . styx . storage . CheckedDatastoreTransaction transaction1 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . newTransaction ( ) ; final com . spotify . styx . storage . CheckedDatastoreTransaction transaction2 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . newTransaction ( ) ; final com . spotify . styx . storage . CheckedDatastoreTransaction transaction3 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . newTransaction ( ) ; com . spotify . styx . storage . DatastoreStorageTransaction storageTransaction1 = new com . spotify . styx . storage . DatastoreStorageTransaction ( transaction1 ) ; com . spotify . styx . storage . DatastoreStorageTransaction storageTransaction2 = new com . spotify . styx . storage . DatastoreStorageTransaction ( transaction2 ) ; com . spotify . styx . storage . DatastoreStorageTransaction storageTransaction3 = new com . spotify . styx . storage . DatastoreStorageTransaction ( transaction3 ) ; com . google . cloud . datastore . KeyFactory keyFactory1 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . newKeyFactory ( ) . setKind ( "MyKind" ) ; com . google . cloud . datastore . Key key1 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . allocateId ( keyFactory1 . newKey ( ) ) ; com . google . cloud . datastore . Entity entity1 = com . google . cloud . datastore . Entity . newBuilder ( key1 ) . set ( "key" , "firstWrite" ) . build ( ) ; transaction1 . put ( entity1 ) ; storageTransaction1 . commit ( ) ; com . google . cloud . datastore . Entity entity1read = transaction3 . get ( key1 ) ; com . google . cloud . datastore . Entity entity1modified = com . google . cloud . datastore . Entity . newBuilder ( key1 ) . set ( "key" , "secondWrite" ) . build ( ) ; transaction2 . put ( entity1modified ) ; storageTransaction2 . commit ( ) ; com . google . cloud . datastore . KeyFactory keyFactory2 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . newKeyFactory ( ) . setKind ( "MyKindCopy" ) ; com . google . cloud . datastore . Key key2 = com . spotify . styx . storage . DatastoreStorageTransactionTest . datastore . allocateId ( keyFactory2 . newKey ( ) ) ; com . google . cloud . datastore . Entity copyOfEntity1 = com . google . cloud . datastore . Entity . newBuilder ( key2 ) . set ( "key" , entity1read . getString ( "key" ) ) . build ( ) ; transaction3 . put ( copyOfEntity1 ) ; try { storageTransaction3 . commit ( ) ; org . junit . Assert . fail ( "Expected<sp>exception!" ) ; } catch ( com . spotify . styx . storage . TransactionException e ) { "<AssertPlaceHolder>" ; transaction3 . rollback ( ) ; } } isConflict ( ) { if ( ( ( getCause ( ) ) != null ) && ( ( getCause ( ) ) instanceof com . google . cloud . datastore . DatastoreException ) ) { com . google . cloud . datastore . DatastoreException datastoreException = ( ( com . google . cloud . datastore . DatastoreException ) ( getCause ( ) ) ) ; return ( datastoreException . getCode ( ) ) == 10 ; } else { return false ; } } | org . junit . Assert . assertTrue ( e . isConflict ( ) ) |
testSizeConstructor ( ) { com . flowpowered . math . matrix . MatrixNf matrix = new com . flowpowered . math . matrix . MatrixNf ( 4 ) ; "<AssertPlaceHolder>" ; } size ( ) { return mat . length ; } | org . junit . Assert . assertEquals ( matrix . size ( ) , 4 ) |
testSerializeNative ( ) { floatSerializer . serializeNative ( com . orientechnologies . common . serialization . types . FloatSerializerTest . OBJECT , stream , 0 ) ; java . lang . Float v = floatSerializer . deserializeNative ( stream , 0 ) ; "<AssertPlaceHolder>" ; } deserializeNative ( byte [ ] , int ) { return java . lang . Float . intBitsToFloat ( com . orientechnologies . common . serialization . types . OFloatSerializer . CONVERTER . getInt ( stream , startPosition , java . nio . ByteOrder . nativeOrder ( ) ) ) ; } | org . junit . Assert . assertEquals ( v , com . orientechnologies . common . serialization . types . FloatSerializerTest . OBJECT ) |
shouldSupportPrefixedPropertiesThatAreNotPartOfGlobalConsumerConfig ( ) { final org . apache . kafka . streams . StreamsConfig streamsConfig = new org . apache . kafka . streams . StreamsConfig ( props ) ; props . put ( org . apache . kafka . streams . StreamsConfig . consumerPrefix ( "interceptor.statsd.host" ) , "host" ) ; final java . util . Map < java . lang . String , java . lang . Object > consumerConfigs = streamsConfig . getGlobalConsumerConfigs ( "clientId" ) ; "<AssertPlaceHolder>" ; } get ( org . apache . kafka . connect . data . Field ) { java . lang . Object val = values [ field . index ( ) ] ; if ( ( val == null ) && ( ( field . schema ( ) . defaultValue ( ) ) != null ) ) { val = field . schema ( ) . defaultValue ( ) ; } return val ; } | org . junit . Assert . assertEquals ( "host" , consumerConfigs . get ( "interceptor.statsd.host" ) ) |
testSpecificSerializeFromSchemaSeveralTimes ( ) { final org . apache . flink . api . java . tuple . Tuple3 < java . lang . Class < ? extends org . apache . avro . specific . SpecificRecord > , org . apache . avro . specific . SpecificRecord , org . apache . flink . types . Row > testData = org . apache . flink . formats . avro . utils . AvroTestUtils . getSpecificTestData ( ) ; final java . lang . String schemaString = testData . f1 . getSchema ( ) . toString ( ) ; final org . apache . flink . formats . avro . AvroRowSerializationSchema serializationSchema = new org . apache . flink . formats . avro . AvroRowSerializationSchema ( schemaString ) ; final org . apache . flink . formats . avro . AvroRowDeserializationSchema deserializationSchema = new org . apache . flink . formats . avro . AvroRowDeserializationSchema ( schemaString ) ; serializationSchema . serialize ( testData . f2 ) ; serializationSchema . serialize ( testData . f2 ) ; final byte [ ] bytes = serializationSchema . serialize ( testData . f2 ) ; final org . apache . flink . types . Row actual = deserializationSchema . deserialize ( bytes ) ; "<AssertPlaceHolder>" ; } deserialize ( org . apache . flink . core . memory . DataInputView ) { final int size = source . readInt ( ) ; final java . util . ArrayList < T > list = new java . util . ArrayList ( size ) ; for ( int i = 0 ; i < size ; i ++ ) { list . add ( elementSerializer . deserialize ( source ) ) ; } return list ; } | org . junit . Assert . assertEquals ( testData . f2 , actual ) |
testGetCurrencySymbolElement ( ) { org . odftoolkit . odfdom . doc . number . OdfCurrencyStyleTest . LOG . info ( "getCurrencySymbolElement" ) ; org . odftoolkit . odfdom . incubator . doc . number . OdfNumberCurrencyStyle instance = new org . odftoolkit . odfdom . incubator . doc . number . OdfNumberCurrencyStyle ( dom , "$" , "$#,##0.00" , "cstyle" ) ; org . odftoolkit . odfdom . dom . element . number . NumberCurrencySymbolElement expResult = new org . odftoolkit . odfdom . dom . element . number . NumberCurrencySymbolElement ( dom ) ; expResult . setTextContent ( "$" ) ; org . odftoolkit . odfdom . dom . element . number . NumberCurrencySymbolElement result = instance . getCurrencySymbolElement ( ) ; "<AssertPlaceHolder>" ; } getTextContent ( ) { return org . odftoolkit . simple . common . TextExtractor . getText ( mTextboxElement ) ; } | org . junit . Assert . assertEquals ( expResult . getTextContent ( ) , result . getTextContent ( ) ) |
testGetTopicOnNullLeaderDoesntThrowNPE ( ) { java . util . Properties props = new java . util . Properties ( ) ; props . put ( KafkaRestConfig . KAFKACLIENT_INIT_TIMEOUT_CONFIG , "100" ) ; io . confluent . kafkarest . KafkaRestConfig config = new io . confluent . kafkarest . KafkaRestConfig ( props ) ; org . apache . kafka . common . Node controller = new org . apache . kafka . common . Node ( 1 , "a" , 1 ) ; org . apache . kafka . clients . admin . MockAdminClient adminClient = new org . apache . kafka . clients . admin . MockAdminClient ( java . util . Arrays . asList ( controller , null ) , controller ) ; org . apache . kafka . common . TopicPartitionInfo partition = new org . apache . kafka . common . TopicPartitionInfo ( 1 , null , java . util . Collections . singletonList ( controller ) , java . util . Collections . singletonList ( controller ) ) ; adminClient . addTopic ( false , "topic" , java . util . Collections . singletonList ( partition ) , new java . util . HashMap < java . lang . String , java . lang . String > ( ) ) ; io . confluent . kafkarest . entities . Topic topic = new io . confluent . kafkarest . AdminClientWrapper ( config , adminClient ) . getTopic ( "topic" ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertEquals ( topic . getName ( ) , "topic" ) |
testParseNegativeAll ( ) { java . lang . String source = ( ( ( "{-1" + ( getDecimalCharacter ( ) ) ) + "2323;<sp>-1" ) + ( getDecimalCharacter ( ) ) ) + "4343}" ; org . hipparchus . geometry . euclidean . twod . Vector2D expected = new org . hipparchus . geometry . euclidean . twod . Vector2D ( ( - 1.2323 ) , ( - 1.4343 ) ) ; org . hipparchus . geometry . euclidean . twod . Vector2D actual = vector2DFormat . parse ( source ) ; "<AssertPlaceHolder>" ; } parse ( java . lang . String ) { java . text . ParsePosition parsePosition = new java . text . ParsePosition ( 0 ) ; org . hipparchus . geometry . euclidean . twod . Vector2D result = parse ( source , parsePosition ) ; if ( ( parsePosition . getIndex ( ) ) == 0 ) { throw new org . hipparchus . exception . MathIllegalStateException ( org . hipparchus . exception . LocalizedCoreFormats . CANNOT_PARSE_AS_TYPE , source , parsePosition . getErrorIndex ( ) , org . hipparchus . geometry . euclidean . twod . Vector2D . class ) ; } return result ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testEndlessLoopProblem ( ) { org . eclipse . swt . graphics . Font font = new org . eclipse . swt . graphics . Font ( display , "Helvetica" , 11 , org . eclipse . swt . SWT . NORMAL ) ; org . eclipse . swt . graphics . Point extent = org . eclipse . rap . rwt . internal . textsize . TextSizeEstimation . textExtent ( font , "Zusatzinfo<sp>(Besuch)" , 100 ) ; "<AssertPlaceHolder>" ; } textExtent ( org . eclipse . swt . graphics . Font , java . lang . String , int ) { int lineCount = 0 ; int maxWidth = 0 ; for ( java . lang . String line : splitNewLines ( string ) ) { lineCount ++ ; int width = org . eclipse . rap . rwt . internal . textsize . TextSizeEstimation . getLineWidth ( line , font ) ; if ( wrapWidth > 0 ) { boolean done = false ; while ( ! done ) { int index = org . eclipse . rap . rwt . internal . textsize . TextSizeEstimation . getLongestMatch ( line , wrapWidth , font ) ; if ( ( index == 0 ) || ( index == ( line . length ( ) ) ) ) { done = true ; } else { java . lang . String substr = line . substring ( 0 , index ) ; width = org . eclipse . rap . rwt . internal . textsize . TextSizeEstimation . getLineWidth ( substr , font ) ; maxWidth = java . lang . Math . max ( maxWidth , width ) ; line = line . substring ( index , line . length ( ) ) ; lineCount ++ ; } } } maxWidth = java . lang . Math . max ( maxWidth , width ) ; } int height = java . lang . Math . round ( ( ( ( org . eclipse . rap . rwt . internal . textsize . TextSizeEstimation . getCharHeight ( font ) ) * 1.25F ) * lineCount ) ) ; return new org . eclipse . swt . graphics . Point ( maxWidth , height ) ; } | org . junit . Assert . assertEquals ( 100 , extent . x ) |
testLongPath ( ) { byte [ ] data = new byte [ 78 ] ; java . nio . ByteBuffer buffer = java . nio . ByteBuffer . wrap ( data ) ; buffer . putShort ( ( ( short ) ( AdobePathSegment . CLOSED_SUBPATH_LENGTH_RECORD ) ) ) ; buffer . putShort ( ( ( short ) ( 1 ) ) ) ; buffer . position ( ( ( buffer . position ( ) ) + 22 ) ) ; buffer . putShort ( ( ( short ) ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_LINKED ) ) ) ; buffer . position ( ( ( buffer . position ( ) ) + 24 ) ) ; buffer . putShort ( ( ( short ) ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_LINKED ) ) ) ; java . awt . geom . Path2D path = new com . twelvemonkeys . imageio . path . AdobePathBuilder ( data ) . path ( ) ; "<AssertPlaceHolder>" ; } path ( ) { java . util . List < java . util . List < com . twelvemonkeys . imageio . path . AdobePathSegment > > subPaths = new java . util . ArrayList < java . util . List < com . twelvemonkeys . imageio . path . AdobePathSegment > > ( ) ; java . util . List < com . twelvemonkeys . imageio . path . AdobePathSegment > currentPath = null ; int currentPathLength = 0 ; com . twelvemonkeys . imageio . path . AdobePathSegment segment ; while ( ( segment = nextSegment ( ) ) != null ) { if ( com . twelvemonkeys . imageio . path . AdobePathBuilder . DEBUG ) { System . out . println ( segment ) ; } if ( ( ( segment . selector ) == ( AdobePathSegment . OPEN_SUBPATH_LENGTH_RECORD ) ) || ( ( segment . selector ) == ( AdobePathSegment . CLOSED_SUBPATH_LENGTH_RECORD ) ) ) { if ( currentPath != null ) { if ( currentPathLength != ( currentPath . size ( ) ) ) { throw new javax . imageio . IIOException ( java . lang . String . format ( "Bad<sp>path,<sp>expected<sp>%d<sp>segments,<sp>found<sp>only<sp>%d" , currentPathLength , currentPath . size ( ) ) ) ; } subPaths . add ( currentPath ) ; } currentPath = new java . util . ArrayList < com . twelvemonkeys . imageio . path . AdobePathSegment > ( segment . length ) ; currentPathLength = segment . length ; } else if ( ( ( ( ( segment . selector ) == ( AdobePathSegment . OPEN_SUBPATH_BEZIER_LINKED ) ) || ( ( segment . selector ) == ( AdobePathSegment . OPEN_SUBPATH_BEZIER_UNLINKED ) ) ) || ( ( segment . selector ) == ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_LINKED ) ) ) || ( ( segment . selector ) == ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_UNLINKED ) ) ) { if ( currentPath == null ) { throw new javax . imageio . IIOException ( "Bad<sp>path,<sp>missing<sp>subpath<sp>length<sp>record" ) ; } if ( ( currentPath . size ( ) ) >= currentPathLength ) { throw new javax . imageio . IIOException ( java . lang . String . format ( "Bad<sp>path,<sp>expected<sp>%d<sp>segments,<sp>found%d" , currentPathLength , currentPath . size ( ) ) ) ; } currentPath . add ( segment ) ; } } if ( currentPath != null ) { if ( currentPathLength != ( currentPath . size ( ) ) ) { throw new javax . imageio . IIOException ( java . lang . String . format ( "Bad<sp>path,<sp>expected<sp>%d<sp>segments,<sp>found<sp>only<sp>%d" , currentPathLength , currentPath . size ( ) ) ) ; } subPaths . add ( currentPath ) ; } return pathToShape ( subPaths ) ; } | org . junit . Assert . assertNotNull ( path ) |
testGetInstaceRole ( ) { com . graphaware . common . policy . role . InstanceRole role = utils . getInstanceRole ( ) ; "<AssertPlaceHolder>" ; } getInstanceRole ( ) { org . neo4j . kernel . impl . factory . OperationalMode operationalMode = getOperationalMode ( ) ; com . graphaware . common . policy . role . InstanceRole res ; switch ( operationalMode ) { case ha : res = getHARole ( ) ; break ; case core : res = getCoreRole ( ) ; break ; case read_replica : res = com . graphaware . common . policy . role . InstanceRole . READ_REPLICA ; break ; default : res = com . graphaware . common . policy . role . InstanceRole . SINGLE ; } return res ; } | org . junit . Assert . assertEquals ( InstanceRole . SINGLE , role ) |
iterator_isEmptyAfterCreation ( ) { "<AssertPlaceHolder>" ; } iterator ( ) { return new com . restfb . Connection . Itr ( this ) ; } | org . junit . Assert . assertFalse ( array . iterator ( ) . hasNext ( ) ) |
testHasCameraService ( ) { com . eclipsesource . tabris . TabrisClient client = new com . eclipsesource . tabris . internal . TabrisClientImpl ( ) ; com . eclipsesource . tabris . camera . Camera camera = client . getService ( com . eclipsesource . tabris . camera . Camera . class ) ; "<AssertPlaceHolder>" ; } getService ( java . lang . Class ) { T result = null ; if ( type == ( com . eclipsesource . tabris . interaction . AppLauncher . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . AppLauncherImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . app . App . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . AppImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . ClientStore . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . ClientStoreImpl . class ) ) ) ; } else if ( ( type == ( com . eclipsesource . tabris . device . ClientDevice . class ) ) || ( type == ( org . eclipse . rap . rwt . client . service . ClientInfo . class ) ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . ClientDeviceImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . push . CloudPush . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . CloudPushImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . camera . Camera . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . CameraImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . camera . PhotoAlbum . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . PhotoAlbumImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . geolocation . Geolocation . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . GeolocationImpl . class ) ) ) ; } else if ( type == ( com . eclipsesource . tabris . print . Printer . class ) ) { result = ( ( T ) ( getSessionInstance ( com . eclipsesource . tabris . internal . PrinterImpl . class ) ) ) ; } else if ( type == ( org . eclipse . rap . rwt . client . service . StartupParameters . class ) ) { result = ( ( T ) ( getSessionInstance ( org . eclipse . rap . rwt . internal . client . StartupParametersImpl . class ) ) ) ; } return result ; } | org . junit . Assert . assertNotNull ( camera ) |
testEntityType ( ) { final org . apache . usergrid . persistence . core . scope . ApplicationScopeImpl applicationScope = new org . apache . usergrid . persistence . core . scope . ApplicationScopeImpl ( new org . apache . usergrid . persistence . model . entity . SimpleId ( "application" ) ) ; final org . apache . usergrid . persistence . model . entity . Id id = new org . apache . usergrid . persistence . model . entity . SimpleId ( "id" ) ; final java . lang . String output = org . apache . usergrid . persistence . index . impl . IndexingUtils . getType ( applicationScope , id ) ; final java . lang . String expected = ( ( ( "appId(" + ( applicationScope . getApplication ( ) . getUuid ( ) ) ) + ",application).entityType(" ) + ( id . getType ( ) ) ) + ")" ; "<AssertPlaceHolder>" ; } getType ( ) { return type ; } | org . junit . Assert . assertEquals ( output , expected ) |
testUpload_forCustomerPage ( ) { bean . setCurrentPMPage ( PriceModelBean . PRICEMODEL_FOR_CUSTOMER ) ; bean . upload ( ) ; verify ( externalCustomerPriceModelCtrl ) . upload ( any ( org . oscm . internal . vo . VOServiceDetails . class ) , any ( org . oscm . internal . vo . VOOrganization . class ) ) ; "<AssertPlaceHolder>" ; } isDirty ( ) { return dirty ; } | org . junit . Assert . assertTrue ( bean . isDirty ( ) ) |
discardOutline ( ) { org . sejda . model . parameter . SplitByEveryXPagesParameters parameters = defaultCommandLine ( ) . withFlag ( "--discardOutline" ) . invokeSejdaConsole ( ) ; "<AssertPlaceHolder>" ; } discardOutline ( ) { return discardOutline ; } | org . junit . Assert . assertTrue ( parameters . discardOutline ( ) ) |
matchesMatchingValues ( ) { matcher . withEntry ( "foo" , "bar" ) ; org . mockserver . matchers . MultiValueMapMatcher mapMatcher = new org . mockserver . matchers . MultiValueMapMatcher ( new org . mockserver . logging . MockServerLogger ( ) , matcher ) ; matched . withEntry ( "foo" , "bar" ) ; "<AssertPlaceHolder>" ; } matches ( org . mockserver . model . HttpRequest , java . lang . String ) { boolean result = false ; if ( ( xpathExpression ) == null ) { mockServerLogger . trace ( context , ( ( ( "Attempting<sp>match<sp>against<sp>null<sp>XPath<sp>Expression<sp>for<sp>[" + matched ) + "]" ) + ( new java . lang . RuntimeException ( ( ( "Attempting<sp>match<sp>against<sp>null<sp>XPath<sp>Expression<sp>for<sp>[" + matched ) + "]" ) ) ) ) ) ; } else if ( matcher . equals ( matched ) ) { result = true ; } else if ( matched != null ) { try { result = ( ( java . lang . Boolean ) ( xpathExpression . evaluate ( stringToXmlDocumentParser . buildDocument ( matched , new org . mockserver . matchers . StringToXmlDocumentParser . ErrorLogger ( ) { @ org . mockserver . matchers . Override public void logError ( final java . lang . String matched , final java . lang . Exception exception ) { mockServerLogger . warn ( context , ( ( ( ( "SAXParseException<sp>while<sp>performing<sp>match<sp>between<sp>[" + ( matcher ) ) + "]<sp>and<sp>[" ) + matched ) + "]" ) , exception ) ; } } ) , XPathConstants . BOOLEAN ) ) ) ; } catch ( java . lang . Exception e ) { mockServerLogger . trace ( context , ( ( ( ( ( "Error<sp>while<sp>matching<sp>xpath<sp>[" + ( matcher ) ) + "]<sp>against<sp>string<sp>[" ) + matched ) + "]<sp>assuming<sp>no<sp>match<sp>-<sp>" ) + ( e . getMessage ( ) ) ) ) ; } } if ( ! result ) { mockServerLogger . trace ( "Failed<sp>to<sp>match<sp>[{}]<sp>with<sp>[{}]" , matched , matcher ) ; } return ( not ) != result ; } | org . junit . Assert . assertTrue ( mapMatcher . matches ( null , matched ) ) |
TestDocumentWithoutInformation ( ) { try { ve = org . apache . pdfbox . preflight . metadata . TestSynchronizedMetadataValidation . sync . validateMetadataSynchronization ( doc , metadata ) ; "<AssertPlaceHolder>" ; } catch ( org . apache . pdfbox . preflight . exception . ValidationException e ) { throw new java . lang . Exception ( e . getMessage ( ) ) ; } } size ( ) { return this . objectPath . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , ve . size ( ) ) |
testOnException ( ) { java . util . concurrent . atomic . AtomicReference < org . slieb . throwables . java . lang . Throwable > reference = new java . util . concurrent . atomic . AtomicReference ( ) ; org . slieb . throwables . java . lang . Exception expected = new org . slieb . throwables . java . lang . Exception ( "expected" ) ; try { org . slieb . throwables . DoubleUnaryOperatorWithThrowable . castDoubleUnaryOperatorWithThrowable ( ( v1 ) -> { throw expected ; } ) . onException ( reference :: set ) . applyAsDouble ( 0 ) ; } catch ( java . lang . Throwable ignored ) { } "<AssertPlaceHolder>" ; } get ( ) { try { return getWithThrowable ( ) ; } catch ( java . lang . RuntimeException | java . lang . Error exception ) { throw exception ; } catch ( final java . lang . Throwable throwable ) { throw new org . slieb . throwables . SuppressedException ( throwable ) ; } } | org . junit . Assert . assertEquals ( expected , reference . get ( ) ) |
testContentSummary ( ) { org . apache . hadoop . hdfs . MiniDFSCluster cluster = null ; final org . apache . hadoop . conf . Configuration conf = org . apache . hadoop . hdfs . web . WebHdfsTestUtil . createConf ( ) ; final org . apache . hadoop . fs . Path path = new org . apache . hadoop . fs . Path ( "/QuotaDir" ) ; try { cluster = new org . apache . hadoop . hdfs . MiniDFSCluster . Builder ( conf ) . numDataNodes ( 0 ) . build ( ) ; final org . apache . hadoop . hdfs . web . WebHdfsFileSystem webHdfs = org . apache . hadoop . hdfs . web . WebHdfsTestUtil . getWebHdfsFileSystem ( conf , WebHdfsConstants . WEBHDFS_SCHEME ) ; final org . apache . hadoop . hdfs . DistributedFileSystem dfs = cluster . getFileSystem ( ) ; dfs . mkdirs ( path ) ; dfs . setQuotaByStorageType ( path , StorageType . DISK , 100000 ) ; org . apache . hadoop . fs . ContentSummary contentSummary = webHdfs . getContentSummary ( path ) ; "<AssertPlaceHolder>" ; } finally { if ( cluster != null ) { cluster . shutdown ( ) ; } } } getTypeQuota ( org . apache . hadoop . fs . StorageType ) { return ( typeQuota ) != null ? typeQuota [ type . ordinal ( ) ] : - 1 ; } | org . junit . Assert . assertTrue ( ( ( contentSummary . getTypeQuota ( StorageType . DISK ) ) == 100000 ) ) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.