input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testTopiPartitionSerializationCompatibility ( ) { java . lang . Object deserializedObject = org . apache . kafka . common . utils . Serializer . deserialize ( fileName ) ; "<AssertPlaceHolder>" ; checkValues ( ( ( org . apache . kafka . common . TopicPartition ) ( deserializedObject ) ) ) ; } deserialize ( java . io . InputStream ) { try ( java . io . ObjectInputStream objectInputStream = new java . io . ObjectInputStream ( inputStream ) ) { return objectInputStream . readObject ( ) ; } }
org . junit . Assert . assertTrue ( ( deserializedObject instanceof org . apache . kafka . common . TopicPartition ) )
testDoMainInstallAndStartBundleWithFileAndLeadingSlash ( ) { org . eclipse . concierge . test . util . SyntheticBundleBuilder builder = org . eclipse . concierge . test . util . SyntheticBundleBuilder . newBuilder ( ) ; builder . bundleSymbolicName ( "-Dorg.eclipse.concierge.log.level=4" 1 ) ; java . io . File f = builder . asFile ( "build/tests/testDoMainInstallAndStartBundleWithFileAndLeadingSlash-0.0.0.jar" ) ; f . deleteOnExit ( ) ; org . eclipse . concierge . Concierge framework = org . eclipse . concierge . Concierge . doMain ( new java . lang . String [ ] { "-Dorg.eclipse.concierge.debug=true" , "-Dorg.eclipse.concierge.log.level=4" 0 , "-Dorg.eclipse.concierge.debug.packages=true" , "-Dorg.eclipse.concierge.debug.services=true" , "-istart" , "file:./" + ( f . toString ( ) ) , "-Dorg.eclipse.concierge.log.enabled=true" , "-Dorg.eclipse.concierge.log.quiet=false" , "-Dorg.eclipse.concierge.log.buffersize=100" , "-Dorg.eclipse.concierge.log.level=4" } ) ; "<AssertPlaceHolder>" ; framework . stop ( ) ; } toString ( ) { return ( "ServiceReference{" + ( service ) ) + "}" ; }
org . junit . Assert . assertNotNull ( framework )
testExportODT ( ) { java . io . File file = new java . io . File ( com . archimatetool . jasperreports . JasperReportsExporterTests . exportFolder , ( ( com . archimatetool . jasperreports . JasperReportsExporterTests . exportFileName ) + ".odt" ) ) ; com . archimatetool . jasperreports . JasperReportsExporterTests . exporter . exportODT ( com . archimatetool . jasperreports . JasperReportsExporterTests . jasperPrint , file ) ; "<AssertPlaceHolder>" ; } exists ( ) { return ( getURL ( ) ) != null ; }
org . junit . Assert . assertTrue ( file . exists ( ) )
testConvert ( ) { java . lang . Long parentId = 1L ; java . lang . Long benefitTypeId = 2L ; java . lang . String abbrName = "fdsf" ; java . util . Date begDate = new java . util . Date ( ) ; java . util . Date endDate = new java . util . Date ( ) ; java . lang . String description = "descriptipp" ; org . lnu . is . domain . benefit . Benefit source = new org . lnu . is . domain . benefit . Benefit ( ) ; source . setAbbrName ( abbrName ) ; org . lnu . is . domain . benefit . BenefitType benefitType = new org . lnu . is . domain . benefit . BenefitType ( ) ; benefitType . setId ( benefitTypeId ) ; source . setBenefitType ( benefitType ) ; source . setBegDate ( begDate ) ; source . setDescription ( description ) ; source . setEndDate ( endDate ) ; source . setName ( source . getName ( ) ) ; org . lnu . is . domain . benefit . Benefit parent = new org . lnu . is . domain . benefit . Benefit ( ) ; parent . setId ( parentId ) ; source . setParent ( parent ) ; org . lnu . is . resource . benefit . BenefitResource expected = new org . lnu . is . resource . benefit . BenefitResource ( ) ; expected . setAbbrName ( abbrName ) ; expected . setBenefitTypeId ( benefitTypeId ) ; expected . setBegDate ( begDate ) ; expected . setDescription ( description ) ; expected . setEndDate ( endDate ) ; expected . setName ( source . getName ( ) ) ; expected . setParentId ( parentId ) ; org . lnu . is . resource . benefit . BenefitResource actual = unit . convert ( source ) ; "<AssertPlaceHolder>" ; } convert ( org . lnu . is . domain . admin . unit . AdminUnit ) { return convert ( source , new org . lnu . is . resource . adminunit . AdminUnitResource ( ) ) ; }
org . junit . Assert . assertEquals ( expected , actual )
testSetVersion ( ) { java . lang . String version = "" ; org . owasp . dependencycheck . xml . pom . Model instance = new org . owasp . dependencycheck . xml . pom . Model ( ) ; instance . setVersion ( version ) ; "<AssertPlaceHolder>" ; } getVersion ( ) { return version ; }
org . junit . Assert . assertNotNull ( instance . getVersion ( ) )
isPropertyAllowed ( ) { io . stardog . starwizard . swagger . AuthParamFilter filter = new io . stardog . starwizard . swagger . AuthParamFilter ( ) ; "<AssertPlaceHolder>" ; } isPropertyAllowed ( io . swagger . models . Model , io . swagger . models . properties . Property , java . lang . String , java . util . Map , java . util . Map , java . util . Map ) { return true ; }
org . junit . Assert . assertTrue ( filter . isPropertyAllowed ( null , null , null , null , null , null ) )
fullUrlStillValidWithParameters ( ) { shortUrl . setFullUrl ( "http://google.com?q=repasky" ) ; constraintViolations = com . repaskys . domain . ShortUrlTest . validator . validate ( shortUrl ) ; "<AssertPlaceHolder>" ; } setFullUrl ( java . lang . String ) { this . fullUrl = fullUrl ; }
org . junit . Assert . assertEquals ( 0 , constraintViolations . size ( ) )
testResourceLimitCheckForUploadedVolume ( ) { doThrow ( new com . cloud . legacymodel . exceptions . ResourceAllocationException ( "primary<sp>storage<sp>resource<sp>limit<sp>check<sp>failed" , Resource . ResourceType . primary_storage ) ) . when ( _svc . _resourceLimitMgr ) . checkResourceLimit ( any ( com . cloud . user . AccountVO . class ) , any ( Resource . ResourceType . class ) , any ( com . cloud . storage . Long . class ) ) ; final com . cloud . vm . UserVmVO vm = org . mockito . Mockito . mock ( com . cloud . vm . UserVmVO . class ) ; final com . cloud . engine . subsystem . api . storage . VolumeInfo volumeToAttach = org . mockito . Mockito . mock ( com . cloud . engine . subsystem . api . storage . VolumeInfo . class ) ; when ( volumeToAttach . getId ( ) ) . thenReturn ( 9L ) ; when ( volumeToAttach . getDataCenterId ( ) ) . thenReturn ( 34L ) ; when ( volumeToAttach . getVolumeType ( ) ) . thenReturn ( VolumeType . DATADISK ) ; when ( volumeToAttach . getInstanceId ( ) ) . thenReturn ( null ) ; when ( _userVmDao . findById ( anyLong ( ) ) ) . thenReturn ( vm ) ; when ( vm . getType ( ) ) . thenReturn ( VirtualMachineType . User ) ; when ( vm . getState ( ) ) . thenReturn ( State . Running ) ; when ( vm . getDataCenterId ( ) ) . thenReturn ( 34L ) ; when ( _svc . _volsDao . findByInstanceAndType ( anyLong ( ) , any ( com . cloud . model . enumeration . VolumeType . class ) ) ) . thenReturn ( new java . util . ArrayList ( 10 ) ) ; when ( _svc . volFactory . getVolume ( 9L ) ) . thenReturn ( volumeToAttach ) ; when ( volumeToAttach . getState ( ) ) . thenReturn ( Volume . State . Uploaded ) ; final com . cloud . dc . DataCenterVO zoneWithDisabledLocalStorage = org . mockito . Mockito . mock ( com . cloud . dc . DataCenterVO . class ) ; when ( _svc . _dcDao . findById ( anyLong ( ) ) ) . thenReturn ( zoneWithDisabledLocalStorage ) ; try { _svc . attachVolumeToVM ( 2L , 9L , null ) ; } catch ( final com . cloud . legacymodel . exceptions . InvalidParameterValueException e ) { "<AssertPlaceHolder>" ; } } getMessage ( ) { return message ; }
org . junit . Assert . assertEquals ( e . getMessage ( ) , "primary<sp>storage<sp>resource<sp>limit<sp>check<sp>failed" )
testModifiedDateIsNull ( ) { final java . lang . Long timestamp = 100000L ; java . text . SimpleDateFormat sdf = new java . text . SimpleDateFormat ( "d<sp>MMM<sp>yyyy<sp>HH:mm:ss<sp>z" ) ; org . pentaho . di . ui . repository . repositoryexplorer . model . UIDatabaseConnection uiconn = new org . pentaho . di . ui . repository . repositoryexplorer . model . UIDatabaseConnection ( ) ; org . pentaho . di . repository . RepositoryElementMetaInterface repoMeta = mock ( org . pentaho . di . repository . RepositoryElementMetaInterface . class ) ; when ( repoMeta . getModifiedDate ( ) ) . thenReturn ( null ) ; uiconn . setRepositoryElementMetaInterface ( repoMeta ) ; "<AssertPlaceHolder>" ; } getDateModified ( ) { java . util . Date dbDate = null ; if ( ( ( repoElementMeta ) != null ) && ( ( repoElementMeta . getModifiedDate ( ) ) != null ) ) { dbDate = repoElementMeta . getModifiedDate ( ) ; } if ( ( ( dbMeta ) != null ) && ( ( dbMeta . getChangedDate ( ) ) != null ) ) { dbDate = dbMeta . getChangedDate ( ) ; } if ( dbDate == null ) { return null ; } java . text . SimpleDateFormat sdf = new java . text . SimpleDateFormat ( "d<sp>MMM<sp>yyyy<sp>HH:mm:ss<sp>z" ) ; return sdf . format ( dbDate ) ; }
org . junit . Assert . assertEquals ( null , uiconn . getDateModified ( ) )
filterWorksWithComparator ( ) { com . metservice . kanban . model . WorkItem workItem1 = new com . metservice . kanban . model . WorkItem ( 1 , featureType ) ; workItem1 . advance ( org . joda . time . LocalDate . parse ( "2012-01-10" ) ) ; com . metservice . kanban . model . WorkItem workItem2 = new com . metservice . kanban . model . WorkItem ( 2 , featureType ) ; workItem2 . advance ( org . joda . time . LocalDate . parse ( "2012-01-06" ) ) ; com . metservice . kanban . model . WorkItem workItem3 = new com . metservice . kanban . model . WorkItem ( 3 , featureType ) ; workItem3 . advance ( org . joda . time . LocalDate . parse ( "2012-01-16" ) ) ; java . util . List < com . metservice . kanban . model . WorkItem > workItems = asList ( workItem1 , workItem2 , workItem3 ) ; com . metservice . kanban . model . KanbanBoardColumn column = new com . metservice . kanban . model . KanbanBoardColumn ( featureType , "phase<sp>1" ) ; com . metservice . kanban . model . KanbanBoardColumnList columnList = new com . metservice . kanban . model . KanbanBoardColumnList ( column ) ; java . util . List < com . metservice . kanban . model . WorkItem > filteredWorkItems = columnList . filter ( workItems , WorkItem . LAST_PHASE_DATE_COMPARATOR ) ; java . util . List < com . metservice . kanban . model . WorkItem > expectedWorkItems = asList ( workItem3 , workItem1 , workItem2 ) ; "<AssertPlaceHolder>" ; } filter ( java . util . List , java . util . Comparator ) { java . util . List < com . metservice . kanban . model . WorkItem > filteredWorkItems = new java . util . ArrayList < com . metservice . kanban . model . WorkItem > ( ) ; java . util . List < com . metservice . kanban . model . WorkItem > workItemsList = new java . util . ArrayList < com . metservice . kanban . model . WorkItem > ( workItems ) ; if ( workItemComparator != null ) { java . util . Collections . sort ( workItemsList , workItemComparator ) ; } for ( com . metservice . kanban . model . WorkItem workItem : workItemsList ) { if ( containsPhase ( workItem . getCurrentPhase ( ) ) ) { filteredWorkItems . add ( workItem ) ; } } return new java . util . ArrayList < com . metservice . kanban . model . WorkItem > ( filteredWorkItems ) ; }
org . junit . Assert . assertThat ( filteredWorkItems , org . hamcrest . core . Is . is ( expectedWorkItems ) )
rotateTest ( ) { com . itextpdf . kernel . geom . AffineTransform rotateOne = com . itextpdf . kernel . geom . AffineTransform . getRotateInstance ( ( ( Math . PI ) / 2 ) ) ; com . itextpdf . kernel . geom . AffineTransform expected = new com . itextpdf . kernel . geom . AffineTransform ( 0 , 1 , ( - 1 ) , 0 , 0 , 0 ) ; "<AssertPlaceHolder>" ; } getRotateInstance ( double ) { com . itextpdf . kernel . geom . AffineTransform t = new com . itextpdf . kernel . geom . AffineTransform ( ) ; t . setToRotation ( angle ) ; return t ; }
org . junit . Assert . assertEquals ( rotateOne , expected )
testRank2OrderByCols ( ) { java . lang . String sqlText = java . lang . String . format ( ( "SELECT<sp>empnum,<sp>dept,<sp>salary,<sp>RANK()<sp>OVER<sp>(ORDER<sp>BY<sp>dept,<sp>salary<sp>desc)<sp>AS<sp>Rank<sp>" + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 2 ) , this . getTableReference ( com . splicemachine . derby . impl . sql . execute . operations . WindowFunctionIT . EMPTAB ) , useSpark ) ; java . sql . ResultSet rs = com . splicemachine . derby . impl . sql . execute . operations . WindowFunctionIT . methodWatcher . executeQuery ( sqlText ) ; java . lang . String expected = "EMPNUM<sp>|DEPT<sp>|SALARY<sp>|RANK<sp>|\n" + ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( "----------------------------\n" + "<sp>10<sp>|<sp>1<sp>|<sp>50000<sp>|<sp>7<sp>|\n" ) + "<sp>20<sp>|<sp>1<sp>|<sp>75000<sp>|<sp>3<sp>|\n" ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" ) + "<sp>40<sp>|<sp>2<sp>|<sp>52000<sp>|<sp>9<sp>|\n" ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 0 ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 7 ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 9 ) + "<sp>55<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>5<sp>|\n" ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 3 ) + "<sp>70<sp>|<sp>1<sp>|<sp>76000<sp>|<sp>2<sp>|\n" ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 4 ) + "<sp>90<sp>|<sp>2<sp>|<sp>51000<sp>|<sp>11<sp>|\n" ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 5 ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 1 ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 6 ) ; "<AssertPlaceHolder>" ; rs . close ( ) ; } toStringUnsorted ( com . splicemachine . homeless . ResultSet ) { return com . splicemachine . homeless . TestUtils . FormattedResult . ResultFactory . convert ( "" , rs , false ) . toString ( ) . trim ( ) ; }
org . junit . Assert . assertEquals ( ( ( "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 8 + sqlText ) + "<sp>30<sp>|<sp>3<sp>|<sp>84000<sp>|<sp>12<sp>|\n" 8 ) , expected , TestUtils . FormattedResult . ResultFactory . toStringUnsorted ( rs ) )
testNonDeterministicEncodingDueToAnOpenBracket ( ) { java . lang . String encoded = this . encodeWithApacheWire ( new java . io . ByteArrayInputStream ( new byte [ ] { '[' } ) ) ; byte [ ] decodedBack = this . decoder . decode ( encoded ) ; "<AssertPlaceHolder>" ; } decode ( java . lang . String ) { if ( ( base64 . length ( ) ) < 30 ) { } else { } java . util . StringTokenizer tok = new java . util . StringTokenizer ( base64 , "<sp>\n\r\t" , false ) ; java . lang . StringBuffer buf = new java . lang . StringBuffer ( base64 . length ( ) ) ; while ( tok . hasMoreElements ( ) ) { buf . append ( tok . nextToken ( ) ) ; } base64 = buf . toString ( ) ; int pad = 0 ; for ( int i = ( base64 . length ( ) ) - 1 ; ( i > 0 ) && ( ( base64 . charAt ( i ) ) == '=' ) ; i -- ) { pad ++ ; } int length = ( ( ( base64 . length ( ) ) / 4 ) * 3 ) - pad ; byte [ ] raw = new byte [ length ] ; for ( int i = 0 , rawIndex = 0 ; i < ( ( base64 . length ( ) ) - 3 ) ; i += 4 , rawIndex += 3 ) { int block = ( ( ( ( org . digidoc4j . ddoc . Base64Util . getValue ( base64 . charAt ( i ) ) ) << 18 ) + ( ( org . digidoc4j . ddoc . Base64Util . getValue ( base64 . charAt ( ( i + 1 ) ) ) ) << 12 ) ) + ( ( org . digidoc4j . ddoc . Base64Util . getValue ( base64 . charAt ( ( i + 2 ) ) ) ) << 6 ) ) + ( org . digidoc4j . ddoc . Base64Util . getValue ( base64 . charAt ( ( i + 3 ) ) ) ) ; for ( int j = 2 ; j >= 0 ; j -- ) { if ( ( rawIndex + j ) < ( raw . length ) ) { raw [ ( rawIndex + j ) ] = ( ( byte ) ( block & 255 ) ) ; } block >>= 8 ; } } return raw ; }
org . junit . Assert . assertArrayEquals ( new byte [ ] { '[' } , decodedBack )
testConvertToAvroStream ( ) { com . datastax . driver . core . ResultSet rs = org . apache . nifi . processors . cassandra . CassandraQueryTestUtil . createMockResultSet ( ) ; java . io . ByteArrayOutputStream baos = new java . io . ByteArrayOutputStream ( ) ; long numberOfRows = org . apache . nifi . processors . cassandra . QueryCassandra . convertToAvroStream ( rs , baos , 0 , null ) ; "<AssertPlaceHolder>" ; } convertToAvroStream ( com . datastax . driver . core . ResultSet , java . io . OutputStream , long , java . util . concurrent . TimeUnit ) { final org . apache . avro . Schema schema = org . apache . nifi . processors . cassandra . QueryCassandra . createSchema ( rs ) ; final org . apache . avro . generic . GenericRecord rec = new org . apache . avro . generic . GenericData . Record ( schema ) ; final org . apache . avro . io . DatumWriter < org . apache . avro . generic . GenericRecord > datumWriter = new org . apache . avro . generic . GenericDatumWriter ( schema ) ; try ( final org . apache . avro . file . DataFileWriter < org . apache . avro . generic . GenericRecord > dataFileWriter = new org . apache . avro . file . DataFileWriter ( datumWriter ) ) { dataFileWriter . create ( schema , outStream ) ; final com . datastax . driver . core . ColumnDefinitions columnDefinitions = rs . getColumnDefinitions ( ) ; long nrOfRows = 0 ; if ( columnDefinitions != null ) { do { int rowsAvailableWithoutFetching = rs . getAvailableWithoutFetching ( ) ; if ( rowsAvailableWithoutFetching == 0 ) { if ( ( timeout <= 0 ) || ( timeUnit == null ) ) { rs . fetchMoreResults ( ) . get ( ) ; } else { rs . fetchMoreResults ( ) . get ( timeout , timeUnit ) ; } } for ( com . datastax . driver . core . Row row : rs ) { for ( int i = 0 ; i < ( columnDefinitions . size ( ) ) ; i ++ ) { final com . datastax . driver . core . DataType dataType = columnDefinitions . getType ( i ) ; if ( row . isNull ( i ) ) { rec . put ( i , null ) ; } else { rec . put ( i , getCassandraObject ( row , i , dataType ) ) ; } } dataFileWriter . append ( rec ) ; nrOfRows += 1 ; } } while ( ! ( rs . isFullyFetched ( ) ) ) ; } return nrOfRows ; } }
org . junit . Assert . assertEquals ( 2 , numberOfRows )
testUpdateBusinessObjectFormat ( ) { org . finra . herd . model . api . xml . BusinessObjectFormat businessObjectFormat = new org . finra . herd . model . api . xml . BusinessObjectFormat ( ID , NAMESPACE , BDEF_NAME , FORMAT_USAGE_CODE , FORMAT_FILE_TYPE_CODE , 1 , true , PARTITION_KEY , FORMAT_DESCRIPTION_2 , NO_FORMAT_DOCUMENT_SCHEMA , NO_FORMAT_DOCUMENT_SCHEMA_URL , NO_ATTRIBUTES , businessObjectFormatServiceTestHelper . getTestAttributeDefinitions ( ) , businessObjectFormatServiceTestHelper . getTestSchema2 ( ) , NO_BUSINESS_OBJECT_FORMAT_PARENTS , NO_BUSINESS_OBJECT_FORMAT_CHILDREN , NO_BUSINESS_OBJECT_FORMAT_EXTERNAL_INTERFACES , NO_RECORD_FLAG_SET , NO_RETENTION_PERIOD_IN_DAYS , NO_RETENTION_TYPE , NO_ALLOW_NON_BACKWARDS_COMPATIBLE_CHANGES_SET ) ; org . finra . herd . model . api . xml . BusinessObjectFormatKey businessObjectFormatKey = new org . finra . herd . model . api . xml . BusinessObjectFormatKey ( NAMESPACE , BDEF_NAME , FORMAT_USAGE_CODE , FORMAT_FILE_TYPE_CODE , INITIAL_FORMAT_VERSION ) ; org . finra . herd . model . api . xml . BusinessObjectFormatUpdateRequest request = businessObjectFormatServiceTestHelper . createBusinessObjectFormatUpdateRequest ( org . finra . herd . rest . FORMAT_DESCRIPTION_2 , org . finra . herd . rest . FORMAT_DOCUMENT_SCHEMA_2 , org . finra . herd . rest . FORMAT_DOCUMENT_SCHEMA_URL , org . finra . herd . rest . NO_ATTRIBUTES , businessObjectFormatServiceTestHelper . getTestSchema2 ( ) ) ; when ( businessObjectFormatService . updateBusinessObjectFormat ( businessObjectFormatKey , request ) ) . thenReturn ( businessObjectFormat ) ; org . finra . herd . model . api . xml . BusinessObjectFormat updatedBusinessObjectFormat = businessObjectFormatRestController . updateBusinessObjectFormat ( org . finra . herd . rest . NAMESPACE , org . finra . herd . rest . BDEF_NAME , org . finra . herd . rest . FORMAT_USAGE_CODE , org . finra . herd . rest . FORMAT_FILE_TYPE_CODE , org . finra . herd . rest . INITIAL_FORMAT_VERSION , request ) ; verify ( businessObjectFormatService ) . updateBusinessObjectFormat ( businessObjectFormatKey , request ) ; verifyNoMoreInteractions ( businessObjectFormatService ) ; "<AssertPlaceHolder>" ; } updateBusinessObjectFormat ( org . finra . herd . model . api . xml . BusinessObjectFormatKey , org . finra . herd . model . api . xml . BusinessObjectFormatUpdateRequest ) { businessObjectFormatHelper . validateBusinessObjectFormatKey ( businessObjectFormatKey ) ; attributeHelper . validateFormatAttributes ( request . getAttributes ( ) ) ; org . finra . herd . model . jpa . BusinessObjectFormatEntity businessObjectFormatEntity = businessObjectFormatDaoHelper . getBusinessObjectFormatEntity ( businessObjectFormatKey ) ; businessObjectFormatEntity . setDescription ( request . getDescription ( ) ) ; businessObjectFormatEntity . setDocumentSchema ( getTrimmedString ( request . getDocumentSchema ( ) ) ) ; businessObjectFormatEntity . setDocumentSchemaUrl ( getTrimmedString ( request . getDocumentSchemaUrl ( ) ) ) ; validateBusinessObjectFormatSchema ( request . getSchema ( ) , businessObjectFormatEntity . getPartitionKey ( ) ) ; updateBusinessObjectFormatAttributesHelper ( businessObjectFormatEntity , request . getAttributes ( ) ) ; org . finra . herd . model . api . xml . BusinessObjectFormat businessObjectFormat = businessObjectFormatHelper . createBusinessObjectFormatFromEntity ( businessObjectFormatEntity ) ; if ( ( ( ( request . getSchema ( ) ) != null ) && ( ! ( request . getSchema ( ) . equals ( businessObjectFormat . getSchema ( ) ) ) ) ) || ( ( ( request . getSchema ( ) ) == null ) && ( ( businessObjectFormat . getSchema ( ) ) != null ) ) ) { clearBusinessObjectFormatSchema ( businessObjectFormatEntity ) ; businessObjectFormatDao . saveAndRefresh ( businessObjectFormatEntity ) ; populateBusinessObjectFormatSchema ( businessObjectFormatEntity , request . getSchema ( ) ) ; } businessObjectFormatEntity = businessObjectFormatDao . saveAndRefresh ( businessObjectFormatEntity ) ; org . finra . herd . service . impl . BusinessObjectFormatServiceImpl . LOGGER . info ( ( "Modify<sp>the<sp>business<sp>object<sp>definition<sp>in<sp>the<sp>search<sp>index<sp>associated<sp>with<sp>the<sp>business<sp>object<sp>definition<sp>format<sp>being<sp>updated." + "<sp>businessObjectDefinitionId=\"{}\",<sp>searchIndexUpdateType=\"{}\"" ) , businessObjectFormatEntity . getBusinessObjectDefinition ( ) . getId ( ) , org . finra . herd . service . impl . SEARCH_INDEX_UPDATE_TYPE_UPDATE ) ; searchIndexUpdateHelper . modifyBusinessObjectDefinitionInSearchIndex ( businessObjectFormatEntity . getBusinessObjectDefinition ( ) , org . finra . herd . service . impl . SEARCH_INDEX_UPDATE_TYPE_UPDATE ) ; messageNotificationEventService . processBusinessObjectFormatVersionChangeNotificationEvent ( businessObjectFormatHelper . getBusinessObjectFormatKey ( businessObjectFormatEntity ) , businessObjectFormatEntity . getBusinessObjectFormatVersion ( ) . toString ( ) ) ; return businessObjectFormatHelper . createBusinessObjectFormatFromEntity ( businessObjectFormatEntity ) ; }
org . junit . Assert . assertEquals ( businessObjectFormat , updatedBusinessObjectFormat )
testDeleteCgroup ( ) { final org . apache . hadoop . yarn . server . nodemanager . util . TestCgroupsLCEResourcesHandlerGPU . MockClock clock = new org . apache . hadoop . yarn . server . nodemanager . util . TestCgroupsLCEResourcesHandlerGPU . MockClock ( ) ; clock . time = java . lang . System . currentTimeMillis ( ) ; org . apache . hadoop . yarn . server . nodemanager . util . CgroupsLCEResourcesHandlerGPU handler = new org . apache . hadoop . yarn . server . nodemanager . util . CgroupsLCEResourcesHandlerGPU ( ) ; handler . setConf ( new org . apache . hadoop . yarn . conf . YarnConfiguration ( ) ) ; handler . initConfig ( ) ; handler . clock = clock ; org . apache . commons . io . FileUtils . deleteQuietly ( org . apache . hadoop . yarn . server . nodemanager . util . TestCgroupsLCEResourcesHandlerGPU . cgroupDir ) ; org . apache . hadoop . yarn . server . nodemanager . util . File tfile = new org . apache . hadoop . yarn . server . nodemanager . util . File ( org . apache . hadoop . yarn . server . nodemanager . util . TestCgroupsLCEResourcesHandlerGPU . cgroupDir . getAbsolutePath ( ) , "tasks" ) ; org . apache . hadoop . yarn . server . nodemanager . util . FileOutputStream fos = org . apache . commons . io . FileUtils . openOutputStream ( tfile ) ; fos . write ( "1234" . getBytes ( ) ) ; fos . close ( ) ; final java . util . concurrent . CountDownLatch latch = new java . util . concurrent . CountDownLatch ( 1 ) ; new java . lang . Thread ( ) { @ org . apache . hadoop . yarn . server . nodemanager . util . Override public void run ( ) { latch . countDown ( ) ; try { java . lang . Thread . sleep ( 200 ) ; } catch ( java . lang . InterruptedException ex ) { } clock . time += org . apache . hadoop . yarn . conf . YarnConfiguration . DEFAULT_NM_LINUX_CONTAINER_CGROUPS_DELETE_TIMEOUT ; } } . start ( ) ; latch . await ( ) ; "<AssertPlaceHolder>" ; org . apache . commons . io . FileUtils . deleteQuietly ( org . apache . hadoop . yarn . server . nodemanager . util . TestCgroupsLCEResourcesHandlerGPU . cgroupDir ) ; } deleteCgroup ( java . lang . String ) { boolean deleted = false ; if ( org . apache . hadoop . yarn . server . nodemanager . util . CgroupsLCEResourcesHandler . LOG . isDebugEnabled ( ) ) { org . apache . hadoop . yarn . server . nodemanager . util . CgroupsLCEResourcesHandler . LOG . debug ( ( "deleteCgroup:<sp>" + cgroupPath ) ) ; } long start = clock . getTime ( ) ; do { try { deleted = checkAndDeleteCgroup ( new java . io . File ( cgroupPath ) ) ; if ( ! deleted ) { java . lang . Thread . sleep ( deleteCgroupDelay ) ; } } catch ( java . lang . InterruptedException ex ) { } } while ( ( ! deleted ) && ( ( ( clock . getTime ( ) ) - start ) < ( deleteCgroupTimeout ) ) ) ; if ( ! deleted ) { org . apache . hadoop . yarn . server . nodemanager . util . CgroupsLCEResourcesHandler . LOG . warn ( ( ( ( ( "Unable<sp>to<sp>delete<sp>cgroup<sp>at:<sp>" + cgroupPath ) + ",<sp>tried<sp>to<sp>delete<sp>for<sp>" ) + ( deleteCgroupTimeout ) ) + "ms" ) ) ; } return deleted ; }
org . junit . Assert . assertFalse ( handler . deleteCgroup ( org . apache . hadoop . yarn . server . nodemanager . util . TestCgroupsLCEResourcesHandlerGPU . cgroupDir . getAbsolutePath ( ) ) )
testUpgradeTemplateVelocityScriptDateUTCFields ( ) { addStructure ( _structureId , DDMStructureConstants . DEFAULT_PARENT_STRUCTURE_ID , DDMStructureConstants . VERSION_DEFAULT , read ( "ddm-structure-date-field.xsd" ) , "xml" ) ; addTemplate ( _templateId , _structureId , null , read ( "ddm-template-with-utc-date-field.vm" ) , "vm" , DDMTemplateConstants . TEMPLATE_TYPE_DISPLAY ) ; _upgradeDynamicDataMapping . upgrade ( ) ; java . lang . String actualDefinition = getTemplateScript ( _templateId ) ; java . lang . String dateFieldName = "$date1" ; com . liferay . petra . string . StringBundler sb = new com . liferay . petra . string . StringBundler ( 3 ) ; sb . append ( "$dateUtil.getDate(" ) ; sb . append ( dateFieldName ) ; sb . append ( "_DateObj,<sp>\"dd<sp>MMM<sp>yyyy<sp>-<sp>HH:mm:ss\",<sp>$locale)" ) ; "<AssertPlaceHolder>" ; } contains ( java . lang . String ) { return com . liferay . segments . service . util . ServiceProps . _instance . _configuration . contains ( key ) ; }
org . junit . Assert . assertTrue ( actualDefinition , actualDefinition . contains ( sb . toString ( ) ) )
testPerLineParserWithHeader ( ) { addPerLineDelimiter ( ) ; addSearch ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 4 , "(\\S*)" ) ; addHeaderDelimiter ( "^THE<sp>HEADER<sp>DELIM.*" ) ; addHeaderValue ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 6 , "header<sp>value:<sp>{text}" ) ; addValue ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 3 , 1 , "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 5 ) ; addValue ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 1 , 2 , "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 5 ) ; com . sonalake . utah . config . FileBuilder file = new com . sonalake . utah . config . FileBuilder ( ) ; file . addLine ( "header<sp>value:<sp>ahoy" ) . addLine ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 0 ) . addLine ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 2 ) . addLine ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" ) ; java . util . List < java . util . Map < java . lang . String , java . lang . String > > results = loadFile ( file ) ; java . util . List < java . util . Map < java . lang . String , java . lang . String > > expectedValues = new java . util . ArrayList < java . util . Map < java . lang . String , java . lang . String > > ( ) { { add ( new java . util . TreeMap < java . lang . String , java . lang . String > ( ) { { put ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 6 , "ahoy" ) ; put ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 3 , "cats" ) ; put ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 1 , "dogs" ) ; } } ) ; add ( new java . util . TreeMap < java . lang . String , java . lang . String > ( ) { { put ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 6 , "ahoy" ) ; put ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 3 , "argy" ) ; put ( "a<sp>line<sp>with<sp>some<sp>argy<sp>and<sp>bargy" 1 , "bargy" ) ; } } ) ; } } ; "<AssertPlaceHolder>" ; } put ( java . lang . String , java . lang . String ) { current ( ) . put ( key , value ) ; return this ; }
org . junit . Assert . assertEquals ( expectedValues , results )
getIndexIdsShouldBeAZeroForSimpleSingletonArray ( ) { array . put ( 0 , array , "a" ) ; "<AssertPlaceHolder>" ; } getIndexIds ( ) { java . lang . Object [ ] ids = getIds ( ) ; java . util . List < java . lang . Integer > indices = new java . util . ArrayList < java . lang . Integer > ( ids . length ) ; for ( java . lang . Object id : ids ) { int int32Id = org . mozilla . javascript . ScriptRuntime . toInt32 ( id ) ; if ( ( int32Id >= 0 ) && ( org . mozilla . javascript . ScriptRuntime . toString ( int32Id ) . equals ( org . mozilla . javascript . ScriptRuntime . toString ( id ) ) ) ) { indices . add ( int32Id ) ; } } return indices . toArray ( new java . lang . Integer [ indices . size ( ) ] ) ; }
org . junit . Assert . assertThat ( array . getIndexIds ( ) , org . hamcrest . core . Is . is ( new java . lang . Integer [ ] { 0 } ) )
shouldCreatePdf ( ) { final com . morgan . design . paf . domain . PafChangeLog log = com . morgan . design . paf . domain . PafChangeLog . createSourceLog ( ) ; log . begin ( ) ; log . finish ( ) ; this . logReport . generate ( log ) ; "<AssertPlaceHolder>" ; new java . io . File ( this . logReport . generateTitle ( ) ) . delete ( ) ; } generateTitle ( ) { final java . text . SimpleDateFormat dateFormat = new java . text . SimpleDateFormat ( this . reportNameDateFormat ) ; final java . lang . String date = dateFormat . format ( new java . util . Date ( ) ) ; return ( ( ( this . reportName ) + "-" ) + date ) + ".pdf" ; }
org . junit . Assert . assertTrue ( new java . io . File ( this . logReport . generateTitle ( ) ) . exists ( ) )
escapeJSON_n ( ) { java . lang . String input = "\n" ; java . lang . String escaped = org . oscm . json . EscapeUtils . escapeJSON ( input ) ; "<AssertPlaceHolder>" ; } escapeJSON ( java . lang . String ) { if ( aText == null ) { return null ; } final java . lang . StringBuilder result = new java . lang . StringBuilder ( ) ; java . text . StringCharacterIterator iterator = new java . text . StringCharacterIterator ( aText ) ; char character = iterator . current ( ) ; while ( character != ( java . text . StringCharacterIterator . DONE ) ) { if ( character == '\"' ) { result . append ( "\\\"" ) ; } else if ( character == '\\' ) { result . append ( "\\\\" ) ; } else if ( character == '/' ) { result . append ( "\\/" ) ; } else if ( character == '\b' ) { result . append ( "\\b" ) ; } else if ( character == '\f' ) { result . append ( "\\f" ) ; } else if ( character == '\n' ) { result . append ( "\\n" ) ; } else if ( character == '\r' ) { result . append ( "\\r" ) ; } else if ( character == '\t' ) { result . append ( "\\t" ) ; } else { result . append ( character ) ; } character = iterator . next ( ) ; } return result . toString ( ) ; }
org . junit . Assert . assertEquals ( "\\n" , escaped )
shouldReportWhenClassIsNotSynthetic ( ) { this . data . access = org . objectweb . asm . Opcodes . ACC_PUBLIC ; final org . pitest . classinfo . ClassInfo testee = new org . pitest . classinfo . ClassInfo ( emptyClassPointer ( ) , emptyClassPointer ( ) , this . data ) ; "<AssertPlaceHolder>" ; } isSynthetic ( ) { return ( ( this . access ) & ( org . objectweb . asm . Opcodes . ACC_SYNTHETIC ) ) != 0 ; }
org . junit . Assert . assertFalse ( testee . isSynthetic ( ) )
testHomePage ( ) { org . apache . wicket . protocol . http . WebApplication app = new com . tddinaction . wicket . normal . MyWicketApp ( ) ; "<AssertPlaceHolder>" ; } getHomePage ( ) { return com . tddinaction . wicket . spring . MyHomePage . class ; }
org . junit . Assert . assertEquals ( com . tddinaction . wicket . normal . MyHomePage . class , app . getHomePage ( ) )
testCreateContextWithFileAtlasMappingFormat ( ) { factory = io . atlasmap . core . DefaultAtlasContextFactory . getInstance ( ) ; factory . init ( ) ; java . io . File file = null ; "<AssertPlaceHolder>" ; } createContext ( io . atlasmap . v2 . AtlasMapping ) { io . atlasmap . core . DefaultAtlasContext context = new io . atlasmap . core . DefaultAtlasContext ( this , mapping ) ; context . init ( ) ; return context ; }
org . junit . Assert . assertNotNull ( factory . createContext ( file ) )
testSearchForExtensionTwoDeepReference ( ) { ca . uhn . fhir . jpa . dao . dstu2 . SearchParameter siblingSp = new ca . uhn . fhir . jpa . dao . dstu2 . SearchParameter ( ) ; siblingSp . setBase ( ResourceTypeEnum . PATIENT ) ; siblingSp . setCode ( "foobar" ) ; siblingSp . setType ( SearchParamTypeEnum . REFERENCE ) ; siblingSp . setXpath ( "Patient.extension('http://acme.org/foo').extension('http://acme.org/bar')" ) ; siblingSp . setXpathUsage ( XPathUsageTypeEnum . NORMAL ) ; siblingSp . setStatus ( ConformanceResourceStatusEnum . ACTIVE ) ; siblingSp . addTarget ( ResourceTypeEnum . APPOINTMENT ) ; mySearchParameterDao . create ( siblingSp , mySrd ) ; mySearchParamRegistry . forceRefresh ( ) ; ca . uhn . fhir . jpa . dao . dstu2 . Appointment apt = new ca . uhn . fhir . jpa . dao . dstu2 . Appointment ( ) ; apt . setStatus ( AppointmentStatusEnum . ARRIVED ) ; org . hl7 . fhir . instance . model . api . IIdType aptId = myAppointmentDao . create ( apt ) . getId ( ) . toUnqualifiedVersionless ( ) ; ca . uhn . fhir . jpa . dao . dstu2 . Patient patient = new ca . uhn . fhir . jpa . dao . dstu2 . Patient ( ) ; patient . addName ( ) . addFamily ( "P2" ) ; ca . uhn . fhir . model . api . ExtensionDt extParent = patient . addUndeclaredExtension ( false , "http://acme.org/foo" ) ; extParent . addUndeclaredExtension ( false , "http://acme.org/bar" ) . setValue ( new ca . uhn . fhir . model . dstu2 . composite . ResourceReferenceDt ( aptId . getValue ( ) ) ) ; org . hl7 . fhir . instance . model . api . IIdType p2id = myPatientDao . create ( patient ) . getId ( ) . toUnqualifiedVersionless ( ) ; ca . uhn . fhir . jpa . searchparam . SearchParameterMap map ; ca . uhn . fhir . rest . api . server . IBundleProvider results ; java . util . List < java . lang . String > foundResources ; map = new ca . uhn . fhir . jpa . searchparam . SearchParameterMap ( ) ; map . add ( "foobar" , new ca . uhn . fhir . jpa . dao . dstu2 . ReferenceParam ( aptId . getValue ( ) ) ) ; results = myPatientDao . search ( map ) ; foundResources = toUnqualifiedVersionlessIdValues ( results ) ; "<AssertPlaceHolder>" ; } contains ( java . lang . Object ) { return data . contains ( o ) ; }
org . junit . Assert . assertThat ( foundResources , contains ( p2id . getValue ( ) ) )
forEachKey ( ) { org . eclipse . collections . api . list . MutableList < java . lang . Integer > collection = Lists . mutable . of ( ) ; org . eclipse . collections . api . map . MutableMap < java . lang . Integer , java . lang . String > map = new org . eclipse . collections . impl . map . fixed . SingletonMap ( 1 , "1" ) ; map . forEachKey ( org . eclipse . collections . impl . block . procedure . CollectionAddProcedure . on ( collection ) ) ; "<AssertPlaceHolder>" ; } newListWith ( E [ ] ) { return new org . eclipse . collections . impl . list . mutable . FastList < > ( elements ) ; }
org . junit . Assert . assertEquals ( org . eclipse . collections . impl . list . mutable . FastList . newListWith ( 1 ) , collection )
g_V_foo_fold_sumXlocalX ( ) { final org . apache . tinkerpop . gremlin . process . traversal . Traversal < org . apache . tinkerpop . gremlin . structure . Vertex , java . lang . Number > traversal = get_g_V_foo_fold_sumXlocalX ( ) ; printTraversalForm ( traversal ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { if ( ! ( this . locked ) ) this . applyStrategies ( ) ; return ( ( this . lastTraverser . bulk ( ) ) > 0L ) || ( this . finalEndStep . hasNext ( ) ) ; }
org . junit . Assert . assertFalse ( traversal . hasNext ( ) )
getHandler ( ) { final io . klerch . alexa . state . model . AlexaStateModel model = new io . klerch . alexa . tellask . dummies . model . AlexaStateModelSample ( ) . withHandler ( handler ) ; final io . klerch . alexa . tellask . model . AlexaIntentModel intentModel = new io . klerch . alexa . tellask . model . AlexaIntentModel ( model ) ; "<AssertPlaceHolder>" ; } getHandler ( ) { return this . model . getHandler ( ) ; }
org . junit . Assert . assertEquals ( handler , intentModel . getHandler ( ) )
test_15_munkelt ( ) { iot . jcypher . query . api . IClause [ ] clauses ; iot . jcypher . query . JcQuery query ; iot . jcypher . query . JcQueryResult result ; java . lang . String cypher ; iot . jcypher . query . values . JcNode n = new iot . jcypher . query . values . JcNode ( "n" ) ; iot . jcypher . query . values . JcNode a = new iot . jcypher . query . values . JcNode ( "a" ) ; iot . jcypher . query . values . JcNode b = new iot . jcypher . query . values . JcNode ( "b" ) ; iot . jcypher . query . values . JcRelation p = new iot . jcypher . query . values . JcRelation ( "p" ) ; iot . jcypher . query . values . JcRelation q = new iot . jcypher . query . values . JcRelation ( "q" ) ; iot . jcypher . query . values . JcRelation r = new iot . jcypher . query . values . JcRelation ( "r" ) ; clauses = new iot . jcypher . query . api . IClause [ ] { iot . jcypher . query . factories . clause . MATCH . node ( a ) . relation ( p ) . out ( ) . node ( b ) , iot . jcypher . query . factories . clause . MATCH . node ( a ) . relation ( q ) . out ( ) . node ( b ) , iot . jcypher . query . factories . clause . WITH . DISTINCT ( ) . value ( a ) , iot . jcypher . query . factories . clause . WITH . value ( b ) , iot . jcypher . query . factories . clause . MATCH . node ( a ) . relation ( r ) . out ( ) . node ( b ) , iot . jcypher . query . factories . clause . CREATE . node ( n ) . label ( "Dummy" ) . property ( "field" ) . value ( r . property ( "field" ) ) , iot . jcypher . query . factories . clause . CREATE . node ( a ) . relation ( ) . out ( ) . type ( "Dummy" ) . node ( n ) , iot . jcypher . query . factories . clause . CREATE . node ( n ) . relation ( ) . out ( ) . type ( "Dummy" ) . node ( b ) , iot . jcypher . query . factories . clause . DO . DELETE ( r ) } ; query = new iot . jcypher . query . JcQuery ( ) ; query . setClauses ( clauses ) ; cypher = print ( clauses , Format . PRETTY_1 ) ; result = test . TempTest . dbAccess . execute ( query ) ; "<AssertPlaceHolder>" ; java . util . List < iot . jcypher . graph . GrNode > as = result . resultOf ( a ) ; java . util . List < iot . jcypher . graph . GrNode > bs = result . resultOf ( b ) ; java . util . List < iot . jcypher . graph . GrRelation > r2s = result . resultOf ( r ) ; return ; } hasErrors ( ) { return ( ! ( this . getGeneralErrors ( ) . isEmpty ( ) ) ) || ( ! ( this . getDBErrors ( ) . isEmpty ( ) ) ) ; }
org . junit . Assert . assertFalse ( result . hasErrors ( ) )
testLongLiteral ( ) { createBigIntegerEntitiesDataSet ( ) ; java . lang . String ejbql = "SELECT<sp>bie<sp>FROM<sp>BigIntegerEntity<sp>bie<sp>WHERE<sp>bie.bigIntegerField<sp>><sp>744073709551615" ; org . apache . cayenne . query . EJBQLQuery query = new org . apache . cayenne . query . EJBQLQuery ( ejbql ) ; java . util . List < org . apache . cayenne . testdo . numeric_types . BigIntegerEntity > result = context . performQuery ( query ) ; "<AssertPlaceHolder>" ; } size ( ) { return neighbors . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , result . size ( ) )
testDeleteLoadBalancer ( ) { try { com . fit2cloud . aliyun . slb . model . request . DeleteLoadBalancerRequest request = new com . fit2cloud . aliyun . slb . model . request . DeleteLoadBalancerRequest ( ) ; request . setLoadBalancerId ( loadBalancerId ) ; com . fit2cloud . aliyun . Response response = client . deleteLoadBalancer ( request ) ; System . out . println ( ( "testDeleteLoadBalancer<sp>::<sp>" + ( new com . google . gson . Gson ( ) . toJson ( response ) ) ) ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; org . junit . Assert . fail ( e . getMessage ( ) ) ; } } deleteLoadBalancer ( com . fit2cloud . aliyun . slb . model . request . DeleteLoadBalancerRequest ) { return gson . fromJson ( request . execute ( "DeleteLoadBalancer" , deleteLoadBalancerRequest . toMap ( ) ) , com . fit2cloud . aliyun . Response . class ) ; }
org . junit . Assert . assertTrue ( true )
getLyricsReturnsV2TagsLyrics ( ) { com . mpatric . mp3agic . ID3v1 id3v1Tag = new com . mpatric . mp3agic . ID3WrapperTest . ID3v1TagForTesting ( ) ; com . mpatric . mp3agic . ID3v2 id3v2Tag = new com . mpatric . mp3agic . ID3WrapperTest . ID3v2TagForTesting ( ) ; id3v2Tag . setLyrics ( "V2<sp>Lyrics" ) ; com . mpatric . mp3agic . ID3Wrapper wrapper = new com . mpatric . mp3agic . ID3Wrapper ( id3v1Tag , id3v2Tag ) ; "<AssertPlaceHolder>" ; } getLyrics ( ) { if ( ( id3v2Tag ) != null ) { return id3v2Tag . getLyrics ( ) ; } else { return null ; } }
org . junit . Assert . assertEquals ( "V2<sp>Lyrics" , wrapper . getLyrics ( ) )
opensValidFile ( ) { com . github . marabou . ui . controller . MainMenuController controllerUnderTest = givenAMainMenuControllerWithMocks ( ) ; controllerUnderTest . audioFileStore = mock ( com . github . marabou . audio . store . AudioFileStore . class ) ; java . io . File file = aValidMockedFile ( ) ; controllerUnderTest . openFile ( file ) ; org . mockito . ArgumentCaptor < com . github . marabou . ui . events . OpenFileEvent > argument = org . mockito . ArgumentCaptor . forClass ( com . github . marabou . ui . events . OpenFileEvent . class ) ; verify ( bus ) . post ( argument . capture ( ) ) ; "<AssertPlaceHolder>" ; } getFile ( ) { return file ; }
org . junit . Assert . assertEquals ( file , argument . getValue ( ) . getFile ( ) )
whenGaurdConditionIsFalse_IsSuccessful ( ) { com . google . common . util . concurrent . Monitor monitor = new com . google . common . util . concurrent . Monitor ( ) ; boolean enteredInCriticalSection = false ; com . google . common . util . concurrent . Monitor . Guard gaurdCondition = monitor . newGuard ( this :: returnFalse ) ; if ( monitor . enterIf ( gaurdCondition ) ) { try { System . out . println ( "Entered<sp>in<sp>critical<sp>section" ) ; enteredInCriticalSection = true ; } finally { monitor . leave ( ) ; } } "<AssertPlaceHolder>" ; }
org . junit . Assert . assertFalse ( enteredInCriticalSection )
validate_shouldPassValidationIfNameHasProperValue ( ) { org . openmrs . EncounterRole role = new org . openmrs . EncounterRole ( ) ; role . setName ( "restraining" ) ; org . springframework . validation . Errors errors = new org . springframework . validation . BindException ( role , "type" ) ; new org . openmrs . validator . RequireNameValidator ( ) . validate ( role , errors ) ; "<AssertPlaceHolder>" ; } hasErrors ( ) { return erroneous ; }
org . junit . Assert . assertFalse ( errors . hasErrors ( ) )
testSendMessageDoesNotWaitForAcceptedOutcome ( io . vertx . ext . unit . TestContext ) { when ( sender . sendQueueFull ( ) ) . thenReturn ( Boolean . FALSE ) ; final org . eclipse . hono . client . MessageSender messageSender = new org . eclipse . hono . client . impl . TelemetrySenderImpl ( config , sender , "tenant" , "telemetry/tenant" , context ) ; final java . util . concurrent . atomic . AtomicReference < io . vertx . core . Handler < io . vertx . proton . ProtonDelivery > > handlerRef = new java . util . concurrent . atomic . AtomicReference ( ) ; doAnswer ( ( invocation ) -> { handlerRef . set ( invocation . getArgument ( 1 ) ) ; return mock ( . class ) ; } ) . when ( sender ) . send ( any ( org . apache . qpid . proton . message . Message . class ) , any ( io . vertx . core . Handler . class ) ) ; final io . vertx . core . Future < io . vertx . proton . ProtonDelivery > result = messageSender . send ( "device" , "some<sp>payload" , "application/text" ) ; final io . vertx . proton . ProtonDelivery rejected = mock ( io . vertx . proton . ProtonDelivery . class ) ; when ( rejected . remotelySettled ( ) ) . thenReturn ( Boolean . TRUE ) ; when ( rejected . getRemoteState ( ) ) . thenReturn ( new org . apache . qpid . proton . amqp . messaging . Rejected ( ) ) ; handlerRef . get ( ) . handle ( rejected ) ; "<AssertPlaceHolder>" ; verify ( sender ) . send ( any ( org . apache . qpid . proton . message . Message . class ) , eq ( handlerRef . get ( ) ) ) ; } succeeded ( ) { return succeeded ; }
org . junit . Assert . assertTrue ( result . succeeded ( ) )
testEmptyEntitlementList ( ) { java . util . List < org . candlepin . dto . api . v1 . EntitlementDTO > ents = poolResource . getPoolEntitlements ( pool1 . getId ( ) , adminPrincipal ) ; "<AssertPlaceHolder>" ; } size ( ) { return virtUuidToConsumerMap . keySet ( ) . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , ents . size ( ) )
testNullHl7Payloads ( ) { java . lang . System . setProperty ( MllpComponent . MLLP_LOG_PHI_PROPERTY , "true" ) ; instance = new org . apache . camel . component . mllp . MllpException ( org . apache . camel . component . mllp . MllpExceptionTest . EXCEPTION_MESSAGE , org . apache . camel . component . mllp . MllpExceptionTest . NULL_BYTE_ARRAY , org . apache . camel . component . mllp . MllpExceptionTest . NULL_BYTE_ARRAY ) ; "<AssertPlaceHolder>" ; } expectedMessage ( java . lang . String , java . lang . String ) { java . lang . StringBuilder expectedMessageBuilder = new java . lang . StringBuilder ( ) ; expectedMessageBuilder . append ( org . apache . camel . component . mllp . MllpExceptionTest . EXCEPTION_MESSAGE ) ; if ( hl7Message != null ) { expectedMessageBuilder . append ( "\n\t{hl7Message<sp>[" ) . append ( hl7Message . length ( ) ) . append ( "]<sp>=<sp>" ) . append ( hl7Message . replaceAll ( "\r" , "<0x0D<sp>CR>" ) . replaceAll ( "\n" , "<0x0A<sp>LF>" ) ) . append ( "}" ) ; } if ( hl7Acknowledgment != null ) { expectedMessageBuilder . append ( "\n\t{hl7Acknowledgement<sp>[" ) . append ( hl7Acknowledgment . length ( ) ) . append ( "]<sp>=<sp>" ) . append ( hl7Acknowledgment . replaceAll ( "\r" , "<0x0D<sp>CR>" ) . replaceAll ( "\n" , "<0x0A<sp>LF>" ) ) . append ( "}" ) ; } return expectedMessageBuilder . toString ( ) ; }
org . junit . Assert . assertEquals ( expectedMessage ( null , null ) , instance . getMessage ( ) )
testSimpleIsRootProject2 ( ) { root . newFile ( "build.gradle" ) ; root . newFile ( "settings.gradle" ) ; org . netbeans . modules . gradle . spi . GradleFiles gf = new org . netbeans . modules . gradle . spi . GradleFiles ( root . getRoot ( ) ) ; "<AssertPlaceHolder>" ; } isRootProject ( ) { return ( ( buildScript ) != null ) && ( rootDir . equals ( projectDir ) ) ; }
org . junit . Assert . assertTrue ( gf . isRootProject ( ) )
returns_greater_than_0_if_first_method_has_more_parameters_than_second ( ) { int result = this . comparator . compare ( many , few ) ; "<AssertPlaceHolder>" ; } compare ( java . lang . reflect . Method , java . lang . reflect . Method ) { java . lang . Integer method1Parameters = method1 . getGenericParameterTypes ( ) . length ; java . lang . Integer method2Parameters = method2 . getGenericParameterTypes ( ) . length ; return method1Parameters . compareTo ( method2Parameters ) ; }
org . junit . Assert . assertTrue ( ( result > 0 ) )
shouldExecuteMathematicalOperations ( ) { com . googlecode . starflow . core . script . spel . SpelScriptEngine scriptEngineImpl = new com . googlecode . starflow . core . script . spel . SpelScriptEngine ( ) ; java . lang . Integer two = ( ( java . lang . Integer ) ( scriptEngineImpl . eval ( "1+1" ) ) ) ; "<AssertPlaceHolder>" ; } eval ( javax . script . ScriptContext ) { return evalExpression ( expression , context ) ; }
org . junit . Assert . assertTrue ( ( two == 2 ) )
testBuildSimpleQuery ( ) { org . opencastproject . workflow . api . WorkflowQuery q = new org . opencastproject . workflow . api . WorkflowQuery ( ) . withMediaPackage ( "123" ) . withSeriesId ( "series1" ) ; java . lang . String solrQuery = dao . createQuery ( q , Permissions . Action . READ . toString ( ) , true ) ; java . lang . String expected = "oc_org:mh_default_org<sp>AND<sp>mediapackageid:123<sp>AND<sp>seriesid:series1" ; "<AssertPlaceHolder>" ; } toString ( ) { return this . getValue ( ) ; }
org . junit . Assert . assertEquals ( expected , solrQuery )
whenIfWorksAsMatcher_thenCorrect ( ) { int input = 3 ; java . lang . String output ; if ( input == 0 ) { output = "zero" ; } if ( input == 1 ) { output = "one" ; } if ( input == 2 ) { output = "two" ; } if ( input == 3 ) { output = "three" ; } else { output = "unknown" ; } "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( "three" , output )
shouldReturnNoChunkedOperationElementsWhenNoElementsInGraph ( ) { final javax . ws . rs . core . Response response = client . executeOperationChunked ( new uk . gov . gchq . gaffer . operation . impl . get . GetAllElements ( ) ) ; final java . util . List < uk . gov . gchq . gaffer . data . element . Element > results = readChunkedElements ( response ) ; "<AssertPlaceHolder>" ; } size ( ) { return map . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , results . size ( ) )
shouldHaveCertificatePolicy ( ) { net . ripe . rpki . commons . crypto . x509cert . X509ResourceCertificate cert = net . ripe . rpki . commons . crypto . x509cert . X509ResourceCertificateTest . createSelfSignedCaResourceCertificate ( ) ; "<AssertPlaceHolder>" ; } getCertificatePolicy ( ) { return net . ripe . rpki . commons . crypto . x509cert . AbstractX509CertificateWrapper . POLICY_OID ; }
org . junit . Assert . assertEquals ( AbstractX509CertificateWrapper . POLICY_OID , cert . getCertificatePolicy ( ) )
testMatrixDeterminant4 ( ) { org . nd4j . OpValidationSuite . ignoreFailing ( ) ; org . nd4j . linalg . factory . Nd4j . getRandom ( ) . setSeed ( 12345 ) ; org . nd4j . linalg . api . ndarray . INDArray in = org . nd4j . linalg . factory . Nd4j . rand ( 4 , 4 ) ; System . out . println ( org . nd4j . autodiff . opvalidation . Arrays . toString ( in . data ( ) . asFloat ( ) ) ) ; org . nd4j . autodiff . samediff . SameDiff sd = org . nd4j . autodiff . samediff . SameDiff . create ( ) ; org . nd4j . autodiff . samediff . SDVariable var = sd . var ( "in" , in ) ; org . nd4j . autodiff . samediff . SDVariable md = sd . f ( ) . matrixDeterminant ( var ) ; double d = new org . apache . commons . math3 . linear . LUDecomposition ( org . nd4j . linalg . checkutil . CheckUtil . convertToApacheMatrix ( in ) ) . getDeterminant ( ) ; java . lang . String err = org . nd4j . autodiff . validation . OpValidation . validate ( new org . nd4j . autodiff . validation . TestCase ( sd ) . expected ( md . getVarName ( ) , org . nd4j . linalg . factory . Nd4j . trueScalar ( d ) ) ) ; "<AssertPlaceHolder>" ; } trueScalar ( java . lang . Number ) { lombok . val ret = org . nd4j . linalg . factory . Nd4j . INSTANCE . trueScalar ( scalar ) ; org . nd4j . linalg . factory . Nd4j . logCreationIfNecessary ( ret ) ; return ret ; }
org . junit . Assert . assertNull ( err )
testAddMany ( ) { java . util . List < org . locationtech . udig . issues . IIssue > newIssues = new java . util . ArrayList < org . locationtech . udig . issues . IIssue > ( ) ; newIssues . add ( org . locationtech . udig . issues . test . IssuesListTestHelper . createFeatureIssue ( "new1" ) ) ; newIssues . add ( org . locationtech . udig . issues . test . IssuesListTestHelper . createFeatureIssue ( "new2" ) ) ; list . addAll ( newIssues ) ; list . refresh ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return ( addedGeoms . size ( ) ) - 1 ; }
org . junit . Assert . assertEquals ( 6 , list . size ( ) )
testLoadCaseInstance_nonExistentCaseInstance ( ) { presenter . loadCaseInstance ( null ) ; "<AssertPlaceHolder>" ; verifyZeroInteractions ( view ) ; verifyZeroInteractions ( caseManagementService ) ; } getCaseRolesAssignments ( ) { return caseRolesAssignments ; }
org . junit . Assert . assertEquals ( 0 , presenter . getCaseRolesAssignments ( ) . size ( ) )
testResolve ( ) { "<AssertPlaceHolder>" ; } toPath ( java . lang . String , java . lang . String [ ] ) { return getFileSystem ( ) . getPath ( string , rest ) ; }
org . junit . Assert . assertEquals ( toPath ( "a" , "b" ) , toPath ( "a" ) . resolve ( "b" ) )
getDiskImageId ( ) { org . apache . commons . configuration . HierarchicalConfiguration configuration = prepareConfiguration ( org . oscm . app . ror . LServerConfigurationTest . DISKIMAGEID , org . oscm . app . ror . LServerConfigurationTest . DISKIMAGEID ) ; lServerConfiguration = new org . oscm . app . ror . data . LServerConfiguration ( configuration ) ; "<AssertPlaceHolder>" ; } getDiskImageId ( ) { return props . getValidatedProperty ( org . oscm . app . iaas . PropertyHandler . DISKIMG_ID ) ; }
org . junit . Assert . assertEquals ( org . oscm . app . ror . LServerConfigurationTest . DISKIMAGEID , lServerConfiguration . getDiskImageId ( ) )
testCheckAndMaintainErrorInMaintenanceFailedMigrations ( ) { when ( vmInstanceDao . listNonMigratingVmsByHostEqualsLastHost ( com . cloud . resource . ResourceManagerImplTest . hostId ) ) . thenReturn ( java . util . Arrays . asList ( vm1 , vm2 ) ) ; boolean enterMaintenanceMode = resourceManager . checkAndMaintain ( com . cloud . resource . ResourceManagerImplTest . hostId ) ; verify ( resourceManager ) . isHostInMaintenance ( host , new java . util . ArrayList ( ) , new java . util . ArrayList ( ) , java . util . Arrays . asList ( vm1 , vm2 ) ) ; verify ( resourceManager ) . setHostIntoErrorInMaintenance ( host , java . util . Arrays . asList ( vm1 , vm2 ) ) ; verify ( resourceManager ) . resourceStateTransitTo ( eq ( host ) , eq ( Event . UnableToMigrate ) , anyLong ( ) ) ; "<AssertPlaceHolder>" ; } eq ( org . apache . http . HttpRequest ) { return argThat ( new com . cloud . utils . rest . HttpRequestMatcher ( request ) ) ; }
org . junit . Assert . assertFalse ( enterMaintenanceMode )
carbon_14 ( ) { org . openscience . cdk . interfaces . IAtom a = g2c . toCDKAtom ( uk . ac . ebi . beam . AtomBuilder . aliphatic ( Element . Carbon ) . hydrogens ( 4 ) . isotope ( 14 ) . build ( ) , 4 ) ; "<AssertPlaceHolder>" ; } getMassNumber ( ) { return java . lang . Integer . valueOf ( 1 ) ; }
org . junit . Assert . assertThat ( a . getMassNumber ( ) , org . hamcrest . CoreMatchers . is ( 14 ) )
restoreHostNameToURL_malformedCurrentURL ( ) { java . lang . String referrerURL = "http://myhost.com/mypage" ; java . lang . String url = "reallyMalformedURL" ; "<AssertPlaceHolder>" ; } restoreHostNameToURL ( java . lang . String , java . lang . String ) { if ( ( referrerURL . startsWith ( "/" ) ) || ( ( referrerURL . length ( ) ) == 0 ) ) { int doubleSlash = url . indexOf ( "//" ) ; int firstSingleSlash = url . indexOf ( "/" , ( doubleSlash + 2 ) ) ; referrerURL = ( url . substring ( 0 , firstSingleSlash ) ) + referrerURL ; } else { try { java . net . URL referrer = new java . net . URL ( referrerURL ) ; java . lang . String referrerHost = referrer . getHost ( ) ; if ( ( referrerHost == null ) || ( ( referrerHost . length ( ) ) == 0 ) ) { java . net . URL currentURL = new java . net . URL ( url ) ; java . lang . String currentHost = currentURL . getHost ( ) ; int doubleSlash = referrerURL . indexOf ( "//" ) ; java . lang . StringBuffer newURLBuf = new java . lang . StringBuffer ( referrerURL ) ; newURLBuf . insert ( ( doubleSlash + 2 ) , currentHost ) ; referrerURL = newURLBuf . toString ( ) ; } } catch ( com . ibm . ws . webcontainer . security . internal . java e ) { if ( ( com . ibm . websphere . ras . TraceComponent . isAnyTracingEnabled ( ) ) && ( com . ibm . ws . webcontainer . security . internal . URLHandler . tc . isDebugEnabled ( ) ) ) { com . ibm . websphere . ras . Tr . debug ( com . ibm . ws . webcontainer . security . internal . URLHandler . tc , "restoreHostNameToURL" , new java . lang . Object [ ] { e } ) ; } } } return referrerURL ; }
org . junit . Assert . assertEquals ( referrerURL , handler . restoreHostNameToURL ( referrerURL , url ) )
testService ( ) { "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( service )
testGetAllUrlsZk ( ) { java . util . Map < java . lang . String , java . lang . String > confOverlay = new java . util . HashMap < java . lang . String , java . lang . String > ( ) ; confOverlay . put ( "hive.server2.zookeeper.publish.configs" , "true" ) ; miniHS2 . start ( confOverlay ) ; java . lang . String directUrl = org . apache . hive . jdbc . HiveConnection . getAllUrls ( miniHS2 . getJdbcURL ( ) ) . get ( 0 ) . getJdbcUriString ( ) ; java . lang . String jdbcUrl = ( ( ( "jdbc:hive2://" + ( miniHS2 . getHost ( ) ) ) + ":" ) + ( miniHS2 . getBinaryPort ( ) ) ) + "/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hs2test;" ; "<AssertPlaceHolder>" ; } getBinaryPort ( ) { return binaryPort ; }
org . junit . Assert . assertEquals ( jdbcUrl , directUrl )
testMultiEmptyString ( ) { java . lang . String input = "" ; java . util . List < java . util . List < com . atilika . kuromoji . naist . jdic . Token > > tokenLists = tokenizer . multiTokenize ( input , 10 , Integer . MAX_VALUE ) ; "<AssertPlaceHolder>" ; } size ( ) { return costs . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , tokenLists . size ( ) )
testGetBusinessObjectData ( ) { org . finra . herd . model . api . xml . BusinessObjectDataKey businessObjectDataKey = new org . finra . herd . model . api . xml . BusinessObjectDataKey ( BDEF_NAMESPACE , BDEF_NAME , FORMAT_USAGE_CODE , FORMAT_FILE_TYPE_CODE , FORMAT_VERSION , PARTITION_VALUE , SUBPARTITION_VALUES , DATA_VERSION ) ; java . lang . String delimitedSubPartitionValues = java . lang . String . join ( "|" , org . finra . herd . rest . SUBPARTITION_VALUES ) ; org . finra . herd . model . api . xml . BusinessObjectData businessObjectData = new org . finra . herd . model . api . xml . BusinessObjectData ( ) ; businessObjectData . setId ( org . finra . herd . rest . ID ) ; when ( herdStringHelper . splitStringWithDefaultDelimiterEscaped ( delimitedSubPartitionValues ) ) . thenReturn ( org . finra . herd . rest . SUBPARTITION_VALUES ) ; when ( businessObjectDataService . getBusinessObjectData ( businessObjectDataKey , org . finra . herd . rest . PARTITION_KEY , org . finra . herd . rest . BDATA_STATUS , org . finra . herd . rest . INCLUDE_BUSINESS_OBJECT_DATA_STATUS_HISTORY , org . finra . herd . rest . INCLUDE_STORAGE_UNIT_STATUS_HISTORY ) ) . thenReturn ( businessObjectData ) ; org . finra . herd . model . api . xml . BusinessObjectData result = businessObjectDataRestController . getBusinessObjectData ( org . finra . herd . rest . BDEF_NAMESPACE , org . finra . herd . rest . BDEF_NAME , org . finra . herd . rest . FORMAT_USAGE_CODE , org . finra . herd . rest . FORMAT_FILE_TYPE_CODE , org . finra . herd . rest . PARTITION_KEY , org . finra . herd . rest . PARTITION_VALUE , delimitedSubPartitionValues , org . finra . herd . rest . FORMAT_VERSION , org . finra . herd . rest . DATA_VERSION , org . finra . herd . rest . BDATA_STATUS , org . finra . herd . rest . INCLUDE_BUSINESS_OBJECT_DATA_STATUS_HISTORY , org . finra . herd . rest . INCLUDE_STORAGE_UNIT_STATUS_HISTORY ) ; verify ( herdStringHelper ) . splitStringWithDefaultDelimiterEscaped ( delimitedSubPartitionValues ) ; verify ( businessObjectDataService ) . getBusinessObjectData ( businessObjectDataKey , org . finra . herd . rest . PARTITION_KEY , org . finra . herd . rest . BDATA_STATUS , org . finra . herd . rest . INCLUDE_BUSINESS_OBJECT_DATA_STATUS_HISTORY , org . finra . herd . rest . INCLUDE_STORAGE_UNIT_STATUS_HISTORY ) ; verifyNoMoreInteractionsHelper ( ) ; "<AssertPlaceHolder>" ; } verifyNoMoreInteractionsHelper ( ) { verifyNoMoreInteractions ( awsHelper , javaPropertiesHelper , retryPolicyFactory , s3Operations ) ; }
org . junit . Assert . assertEquals ( businessObjectData , result )
converterCannotInitializeWithCommandAndAttribute ( ) { channelProperties . put ( "zigbee_shortpress_cluster_id" , "0x0008" ) ; channelProperties . put ( "zigbee_shortpress_command_id" , "0xabc" ) ; channelProperties . put ( "zigbee_shortpress_attribute_id" , "1" ) ; channelProperties . put ( "zigbee_shortpress_attribute_value" , "2" ) ; mockCluster ( 8 ) ; boolean initResult = converter . initializeConverter ( ) ; "<AssertPlaceHolder>" ; } initializeConverter ( ) { cluster = ( ( com . zsmartsystems . zigbee . zcl . clusters . ZclRelativeHumidityMeasurementCluster ) ( endpoint . getInputCluster ( ZclRelativeHumidityMeasurementCluster . CLUSTER_ID ) ) ) ; if ( ( cluster ) == null ) { logger . error ( "{}:<sp>Error<sp>opening<sp>device<sp>relative<sp>humidity<sp>measurement<sp>cluster" , endpoint . getIeeeAddress ( ) ) ; return false ; } bind ( cluster ) ; cluster . addAttributeListener ( this ) ; try { com . zsmartsystems . zigbee . CommandResult response = cluster . setMeasuredValueReporting ( 1 , org . openhab . binding . zigbee . internal . converter . REPORTING_PERIOD_DEFAULT_MAX , 0.1 ) . get ( ) ; handleReportingResponse ( response , org . openhab . binding . zigbee . internal . converter . POLLING_PERIOD_DEFAULT , org . openhab . binding . zigbee . internal . converter . REPORTING_PERIOD_DEFAULT_MAX ) ; } catch ( java . lang . InterruptedException | java . util . concurrent . ExecutionException e ) { logger . error ( "{}:<sp>Exception<sp>setting<sp>reporting<sp>" , endpoint . getIeeeAddress ( ) , e ) ; pollingPeriod = POLLING_PERIOD_HIGH ; } return true ; }
org . junit . Assert . assertFalse ( initResult )
testIndexedDotProduct ( ) { mikera . vectorz . Vector v1 = mikera . vectorz . Vector . of ( 0 , 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 ) ; mikera . vectorz . Vector v2 = mikera . vectorz . Vector . of ( 1 , 2 , 3 ) ; mikera . indexz . Index ix = mikera . indexz . Index . of ( 2 , 7 , 4 ) ; "<AssertPlaceHolder>" ; } dotProduct ( double [ ] , int ) { return ( ( ( ( x ) * ( as [ offset ] ) ) + ( ( y ) * ( as [ ( offset + 1 ) ] ) ) ) + ( ( z ) * ( as [ ( offset + 2 ) ] ) ) ) + ( ( t ) * ( as [ ( offset + 3 ) ] ) ) ; }
org . junit . Assert . assertEquals ( ( ( ( 1 * 2 ) + ( 2 * 7 ) ) + ( 3 * 4 ) ) , v1 . dotProduct ( v2 , ix ) , 0.0 )
ignoreStatic ( ) { net . openhft . chronicle . wire . Wire wire = new net . openhft . chronicle . wire . TextWire ( net . openhft . chronicle . bytes . Bytes . elasticHeapByteBuffer ( 256 ) ) ; net . openhft . chronicle . core . io . Closeable writer = wire . methodWriter ( net . openhft . chronicle . core . io . Closeable . class ) ; net . openhft . chronicle . core . io . Closeable . closeQuietly ( writer ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( "NestedClass{" + "text='" ) + ( text ) ) + '\'' ) + ",<sp>number=" ) + ( number ) ) + '}' ; }
org . junit . Assert . assertEquals ( "" , wire . toString ( ) )
nietValideDoorSoortSynchronisatieNull ( ) { final nl . bzk . brp . model . internbericht . ProtocolleringOpdracht protocolleringOpdracht = maakProtocolleringOpdracht ( SoortDienst . ATTENDERING , datumMaterieelSelectie , datumAanvangMaterielePeriode , datumEindeMaterielePeriode , datumTijdAanvangFormelePeriode , datumTijdEindeFormelePeriode , null ) ; org . springframework . test . util . ReflectionTestUtils . setField ( protocolleringOpdracht . getLevering ( ) , "soortSynchronisatie" , null ) ; "<AssertPlaceHolder>" ; } isValide ( ) { boolean resultaat ; if ( ( levering ) == null ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "Levering<sp>dient<sp>gevuld<sp>te<sp>zijn." ) ; } else if ( ( ( personen ) == null ) || ( personen . isEmpty ( ) ) ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "Personen<sp>dient<sp>gevuld<sp>te<sp>zijn." ) ; } else if ( ( levering . getToegangLeveringsautorisatieId ( ) ) == null ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "ToegangAbonnementId<sp>dient<sp>gevuld<sp>te<sp>zijn." ) ; } else if ( ( levering . getDienstId ( ) ) == null ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "DienstId<sp>dient<sp>gevuld<sp>te<sp>zijn." ) ; } else if ( ( ( levering . getDatumTijdKlaarzettenLevering ( ) ) == null ) || ( levering . getDatumTijdKlaarzettenLevering ( ) . heeftGeenWaarde ( ) ) ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "DatumTijdKlaarzettenLevering<sp>dient<sp>gevuld<sp>te<sp>zijn." ) ; } else if ( ( getSoortDienst ( ) ) == null ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "Soort<sp>dienst<sp>gevuld<sp>te<sp>zijn." ) ; } else if ( ( nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . SOORTDIENSTEN_MET_SOORT_SYNCHRONISATIE_VERPLICHT . contains ( getSoortDienst ( ) ) ) && ( ( ( levering . getSoortSynchronisatie ( ) ) == null ) || ( levering . getSoortSynchronisatie ( ) . heeftGeenWaarde ( ) ) ) ) { resultaat = false ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( "ToegangAbonnementId<sp>dient<sp>gevuld<sp>te<sp>zijn." 0 , getSoortDienst ( ) ) ; } else { switch ( soortDienst ) { case ATTENDERING : case MUTATIELEVERING_OP_BASIS_VAN_DOELBINDING : resultaat = isValideAttenderingOfMutatieLeveringDoelbinding ( ) ; break ; case GEEF_DETAILS_PERSOON : case GEEF_DETAILS_PERSOON_BULK : resultaat = isValideGeefDetailsPersoon ( getHistorievorm ( ) ) ; break ; case MUTATIELEVERING_OP_BASIS_VAN_AFNEMERINDICATIE : case PLAATSEN_AFNEMERINDICATIE : case VERWIJDEREN_AFNEMERINDICATIE : resultaat = isValideAfnemerindicatie ( ) ; break ; case SYNCHRONISATIE_PERSOON : resultaat = isValideSynchronisatiePersoon ( ) ; break ; case GEEF_MEDEBEWONERS_VAN_PERSOON : resultaat = isValideGeefMedebewonersVanPersoon ( ) ; break ; default : final java . lang . String foutmelding = "Voor<sp>deze<sp>catalogusoptie<sp>is<sp>geen<sp>protocollering<sp>validatie<sp>ingesteld:<sp>" + ( soortDienst ) ; nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . error ( foutmelding ) ; throw new java . lang . IllegalArgumentException ( foutmelding ) ; } if ( ! resultaat ) { nl . bzk . brp . model . internbericht . ProtocolleringOpdracht . LOGGER . debug ( ( "De<sp>protocollering<sp>is<sp>niet<sp>valide<sp>voor<sp>de<sp>catalogusoptie:<sp>{},<sp>" + ( ( "datum<sp>materieel<sp>selectie:<sp>{},<sp>datum<sp>aanvang<sp>materiele<sp>periode:<sp>{},<sp>" + "datum<sp>einde<sp>materiele<sp>periode:<sp>{},<sp>datum<sp>tijd<sp>aanv<sp>form<sp>periode:<sp>{},<sp>" ) + "ToegangAbonnementId<sp>dient<sp>gevuld<sp>te<sp>zijn." 1 ) ) , soortDienst , levering . getDatumMaterieelSelectie ( ) , levering . getDatumAanvangMaterielePeriodeResultaat ( ) , levering . getDatumEindeMaterielePeriodeResultaat ( ) , levering . getDatumTijdAanvangFormelePeriodeResultaat ( ) , levering . getDatumTijdEindeFormelePeriodeResultaat ( ) , historievorm ) ; } } return resultaat ; }
org . junit . Assert . assertFalse ( protocolleringOpdracht . isValide ( ) )
testFetchByPrimaryKeysWithNoPrimaryKeys ( ) { java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; java . util . Map < java . io . Serializable , com . liferay . social . kernel . model . SocialActivitySet > socialActivitySets = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( socialActivitySets . isEmpty ( ) )
testIncreasedByConditionWithRampPV ( ) { final org . csstudio . scan . device . PVDevice device = getDemoDevice ( ) ; device . start ( ) ; device . write ( 1.0 ) ; final java . lang . Thread ramp = new java . lang . Thread ( new java . lang . Runnable ( ) { @ org . csstudio . scan . Override public void run ( ) { try { for ( int i = 1 ; i <= 5 ; ++ i ) { java . lang . Thread . sleep ( 500L ) ; System . out . println ( ( "Setting<sp>value<sp>to<sp>" + i ) ) ; device . write ( java . lang . Double . valueOf ( i ) ) ; } } catch ( java . lang . Exception ex ) { ex . printStackTrace ( ) ; } } } , "RampValues" ) ; ramp . start ( ) ; final org . csstudio . scan . condition . NumericValueCondition condition = new org . csstudio . scan . condition . NumericValueCondition ( device , org . csstudio . scan . command . Comparison . INCREASE_BY , 3.0 , 0.0 , null ) ; "<AssertPlaceHolder>" ; System . out . println ( ( "Initial<sp>value:<sp>" + ( org . csstudio . scan . device . VTypeHelper . toDouble ( device . read ( ) ) ) ) ) ; condition . await ( ) ; System . out . println ( "Value<sp>increased<sp>by<sp>3!" ) ; ramp . join ( ) ; device . stop ( ) ; } isConditionMet ( ) { final double value = org . csstudio . scan . device . VTypeHelper . toDouble ( device . read ( ) ) ; switch ( comparison ) { case EQUALS : return ( java . lang . Math . abs ( ( ( desired_value ) - value ) ) ) <= ( tolerance ) ; case AT_LEAST : return value >= ( desired_value ) ; case ABOVE : return value > ( desired_value ) ; case AT_MOST : return value <= ( desired_value ) ; case BELOW : return value < ( desired_value ) ; case INCREASE_BY : return value >= ( ( initial_value ) + ( desired_value ) ) ; case DECREASE_BY : return value <= ( ( initial_value ) - ( desired_value ) ) ; default : throw new java . lang . Error ( ( "Condition<sp>not<sp>implemented:<sp>" + ( comparison ) ) ) ; } }
org . junit . Assert . assertThat ( condition . isConditionMet ( ) , org . hamcrest . CoreMatchers . equalTo ( false ) )
testSearchBusinessObjectDefinitionsMissingOptionalParams ( ) { java . util . Set < org . finra . herd . model . api . xml . BusinessObjectDefinition > expectedBusinessObjectDefinitions = setUpTestEntitiesForSearchTesting ( ) ; for ( org . finra . herd . model . api . xml . BusinessObjectDefinition businessObjectDefinition : expectedBusinessObjectDefinitions ) { businessObjectDefinition . setDisplayName ( null ) ; businessObjectDefinition . setShortDescription ( null ) ; businessObjectDefinition . setDataProviderName ( null ) ; } org . finra . herd . model . api . xml . BusinessObjectDefinitionSearchResponse searchResponse = businessObjectDefinitionService . searchBusinessObjectDefinitions ( new org . finra . herd . model . api . xml . BusinessObjectDefinitionSearchRequest ( java . util . Arrays . asList ( new org . finra . herd . model . api . xml . BusinessObjectDefinitionSearchFilter ( NO_EXCLUSION_SEARCH_FILTER , java . util . Arrays . asList ( new org . finra . herd . model . api . xml . BusinessObjectDefinitionSearchKey ( new org . finra . herd . model . api . xml . TagKey ( TAG_TYPE , TAG_CODE ) , INCLUDE_TAG_HIERARCHY ) ) ) ) ) , com . google . common . collect . Sets . newHashSet ( org . finra . herd . service . BLANK_TEXT ) ) ; java . util . Set < org . finra . herd . model . api . xml . BusinessObjectDefinition > actualBusinessObjectDefinitions = new java . util . HashSet ( searchResponse . getBusinessObjectDefinitions ( ) ) ; "<AssertPlaceHolder>" ; } getBusinessObjectDefinitions ( ) { return businessObjectDefinitionService . getBusinessObjectDefinitions ( ) ; }
org . junit . Assert . assertEquals ( actualBusinessObjectDefinitions , expectedBusinessObjectDefinitions )
testBadRGBColourConstructorArgumentLength ( ) { byte [ ] rgb = new byte [ 3 ] ; org . codice . imaging . nitf . core . RGBColour goodColour = new org . codice . imaging . nitf . core . impl . RGBColourImpl ( rgb ) ; "<AssertPlaceHolder>" ; byte [ ] shortArray = new byte [ 2 ] ; exception . expect ( org . codice . imaging . nitf . core . common . NitfFormatException . class ) ; exception . expectMessage ( "Incorrect<sp>number<sp>of<sp>bytes<sp>in<sp>RGB<sp>constructor<sp>array" ) ; org . codice . imaging . nitf . core . RGBColour failColour = new org . codice . imaging . nitf . core . impl . RGBColourImpl ( shortArray ) ; }
org . junit . Assert . assertNotNull ( goodColour )
testRecoverFromParseError ( ) { org . openrdf . sail . Sail sail = new org . openrdf . sail . memory . MemoryStore ( ) ; sail . initialize ( ) ; java . lang . String bad = "bad" ; java . lang . String good = "@prefix<sp>foo:<sp><http://example.org/foo#>.\n" + "foo:a<sp>foo:b<sp>foo:c." ; java . io . InputStream is = new java . io . ByteArrayInputStream ( bad . getBytes ( ) ) ; try { add ( sail , is , "" , RDFFormat . TURTLE ) ; } catch ( java . lang . Exception ignored ) { } finally { is . close ( ) ; } try { is = new java . io . ByteArrayInputStream ( good . getBytes ( ) ) ; add ( sail , is , "" , RDFFormat . TURTLE ) ; } catch ( java . lang . Exception ignored ) { } is . close ( ) ; org . openrdf . sail . SailConnection sc = sail . getConnection ( ) ; try { sc . begin ( ) ; int count = net . fortytwo . ripple . model . impl . sesame . SesameTest . countStatements ( sc , null ) ; "<AssertPlaceHolder>" ; } finally { sc . rollback ( ) ; sc . close ( ) ; } } countStatements ( org . openrdf . sail . SailConnection , org . openrdf . model . IRI ) { int count = 0 ; info . aduna . iteration . CloseableIteration < ? extends org . openrdf . model . Statement , org . openrdf . sail . SailException > stmtIter = ( null == context ) ? sc . getStatements ( null , null , null , false ) : sc . getStatements ( null , null , null , false , context ) ; while ( stmtIter . hasNext ( ) ) { stmtIter . next ( ) ; count ++ ; } stmtIter . close ( ) ; return count ; }
org . junit . Assert . assertEquals ( 1 , count )
deveObterTempoMedioComoFoiSetado ( ) { final com . fincatto . documentofiscal . nfe310 . classes . statusservico . consulta . NFStatusServicoConsultaRetorno consultaRetorno = new com . fincatto . documentofiscal . nfe310 . classes . statusservico . consulta . NFStatusServicoConsultaRetorno ( ) ; final java . lang . String tempoMedio = "10" ; consultaRetorno . setTempoMedio ( tempoMedio ) ; "<AssertPlaceHolder>" ; } getTempoMedio ( ) { return this . tempoMedio ; }
org . junit . Assert . assertEquals ( tempoMedio , consultaRetorno . getTempoMedio ( ) )
updateGuest ( ) { org . candlepin . dto . api . v1 . GuestIdDTO guest = org . candlepin . test . TestUtil . createGuestIdDTO ( "some_guest" ) ; org . candlepin . model . GuestId guestEnt = new org . candlepin . model . GuestId ( ) ; guestEnt . setId ( "some_id" ) ; guestIdResource . updateGuest ( consumer . getUuid ( ) , guest . getGuestId ( ) , guest ) ; when ( guestIdCurator . findByGuestIdAndOrg ( anyString ( ) , any ( java . lang . String . class ) ) ) . thenReturn ( guestEnt ) ; org . mockito . ArgumentCaptor < org . candlepin . model . GuestId > guestCaptor = org . mockito . ArgumentCaptor . forClass ( org . candlepin . model . GuestId . class ) ; org . candlepin . resource . Mockito . verify ( guestIdCurator , org . candlepin . resource . Mockito . times ( 1 ) ) . merge ( guestCaptor . capture ( ) ) ; org . candlepin . model . GuestId result = guestCaptor . getValue ( ) ; "<AssertPlaceHolder>" ; } getConsumer ( ) { return consumer ; }
org . junit . Assert . assertEquals ( consumer , result . getConsumer ( ) )
testSetDimension1Success ( ) { org . goodoldai . jeff . explanation . data . Dimension dim = new org . goodoldai . jeff . explanation . data . Dimension ( "Miles" ) ; three . setDimension1 ( dim ) ; "<AssertPlaceHolder>" ; } getDimension1 ( ) { return dimension1 ; }
org . junit . Assert . assertEquals ( dim , three . getDimension1 ( ) )
testGetLayerGridNotAttachedToLayer ( ) { "<AssertPlaceHolder>" ; } getLayer ( ) { return gridLayer ; }
org . junit . Assert . assertEquals ( gridLayer , grid . getLayer ( ) )
non_strict_with_passed_scenarios ( ) { createNonStrictExitStatus ( ) ; bus . send ( testCaseFinishedWithStatus ( Result . Type . PASSED ) ) ; "<AssertPlaceHolder>" ; } exitStatus ( ) { if ( results . isEmpty ( ) ) { return cucumber . runtime . ExitStatus . DEFAULT ; } if ( runtimeOptions . isWip ( ) ) { return min ( results , cucumber . runtime . SEVERITY ) . is ( Result . Type . PASSED ) ? cucumber . runtime . ExitStatus . ERRORS : cucumber . runtime . ExitStatus . DEFAULT ; } return max ( results , cucumber . runtime . SEVERITY ) . isOk ( runtimeOptions . isStrict ( ) ) ? cucumber . runtime . ExitStatus . DEFAULT : cucumber . runtime . ExitStatus . ERRORS ; }
org . junit . Assert . assertEquals ( 0 , exitStatus . exitStatus ( ) )
testSerialization ( ) { org . jfree . chart . needle . LineNeedle n1 = new org . jfree . chart . needle . LineNeedle ( ) ; org . jfree . chart . needle . LineNeedle n2 = ( ( org . jfree . chart . needle . LineNeedle ) ( org . jfree . chart . TestUtils . serialised ( n1 ) ) ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( obj == ( this ) ) { return true ; } if ( ! ( obj instanceof org . jfree . data . xy . YWithXInterval ) ) { return false ; } org . jfree . data . xy . YWithXInterval that = ( ( org . jfree . data . xy . YWithXInterval ) ( obj ) ) ; if ( ( this . y ) != ( that . y ) ) { return false ; } if ( ( this . xLow ) != ( that . xLow ) ) { return false ; } if ( ( this . xHigh ) != ( that . xHigh ) ) { return false ; } return true ; }
org . junit . Assert . assertTrue ( n1 . equals ( n2 ) )
write ( ) { new com . restfb . json . JsonNumber ( "23" ) . write ( writer ) ; "<AssertPlaceHolder>" ; } toString ( ) { return com . restfb . util . ReflectionUtils . toString ( this ) ; }
org . junit . Assert . assertEquals ( "23" , output . toString ( ) )
checkToBigIntervalCausingOverflow ( ) { build . pluto . test . dependency . MockRemoteRequirement req = new build . pluto . test . dependency . MockRemoteRequirement ( tsPath , Long . MAX_VALUE ) ; req . setIsRemoteAccessible ( true ) ; req . setIsLocalAvailable ( true ) ; boolean consistency = writeTSOnce ( req , 13000L , false ) ; long contentOfFile = readTimestampFromFile ( tsPath ) ; "<AssertPlaceHolder>" ; } readTimestampFromFile ( java . io . File ) { try { java . lang . String persistentPathContent = org . sugarj . common . FileCommands . readFileAsString ( file ) ; return java . lang . Long . parseLong ( persistentPathContent . replace ( "\n" , "" ) ) ; } catch ( java . io . IOException e ) { org . junit . Assert . fail ( "File<sp>that<sp>contains<sp>timestamp<sp>could<sp>not<sp>be<sp>read" ) ; } catch ( java . lang . NumberFormatException e ) { org . junit . Assert . fail ( "Timestamp<sp>was<sp>not<sp>in<sp>the<sp>correct<sp>format" ) ; } return - 1L ; }
org . junit . Assert . assertEquals ( true , consistency )
testFilterConfigurationLocation_Parameter ( ) { com . mockrunner . mock . web . MockFilterConfig config1 = new com . mockrunner . mock . web . MockFilterConfig ( ) ; config1 . setFilterName ( "cayenne-x" ) ; config1 . setInitParameter ( WebConfiguration . CONFIGURATION_LOCATION_PARAMETER , "cayenne-y.xml" ) ; org . apache . cayenne . configuration . web . WebConfiguration configuration1 = new org . apache . cayenne . configuration . web . WebConfiguration ( config1 ) ; "<AssertPlaceHolder>" ; } getConfigurationLocation ( ) { java . lang . String configurationLocation = configuration . getInitParameter ( org . apache . cayenne . configuration . web . WebConfiguration . CONFIGURATION_LOCATION_PARAMETER ) ; if ( configurationLocation != null ) { return configurationLocation ; } java . lang . String name = configuration . getFilterName ( ) ; if ( name == null ) { return null ; } if ( ! ( name . endsWith ( ".xml" ) ) ) { name = name + ".xml" ; } return name ; }
org . junit . Assert . assertEquals ( "cayenne-y.xml" , configuration1 . getConfigurationLocation ( ) )
testInstantiatesRowIfFirstCellOfExampleContainsFullyQualifiedClassName ( ) { info . novatec . testit . livingdoc . util . Rows rows = info . novatec . testit . livingdoc . util . Rows . parse ( ( ( "[" + ( info . novatec . testit . livingdoc . interpreter . flow . dowith . AcceptRow . class . getName ( ) ) ) + "]" ) ) ; info . novatec . testit . livingdoc . interpreter . flow . Row row = selector . select ( rows ) ; "<AssertPlaceHolder>" ; } instanceOf ( java . lang . Class ) { return new info . novatec . testit . livingdoc . expectation . IsInstanceExpectation ( c ) ; }
org . junit . Assert . assertThat ( row , org . hamcrest . Matchers . is ( org . hamcrest . Matchers . instanceOf ( info . novatec . testit . livingdoc . interpreter . flow . dowith . AcceptRow . class ) ) )
testIsGatewayUsingDiskStoreWithNullGatewayQueueAttributes ( ) { final com . gemstone . gemfire . cache . util . Gateway mockGateway = mockContext . mock ( com . gemstone . gemfire . cache . util . Gateway . class , "Gateway" ) ; final com . gemstone . gemfire . cache . DiskStore mockDiskStore = mockContext . mock ( com . gemstone . gemfire . cache . DiskStore . class , "DiskStore" ) ; mockContext . checking ( new org . jmock . Expectations ( ) { { oneOf ( mockGateway ) . getQueueAttributes ( ) ; will ( returnValue ( null ) ) ; oneOf ( mockDiskStore ) . getName ( ) ; will ( returnValue ( DiskStoreDetails . DEFAULT_DISK_STORE_NAME ) ) ; } } ) ; final com . gemstone . gemfire . management . internal . cli . functions . DescribeDiskStoreFunction function = createDescribeDiskStoreFunction ( null ) ; "<AssertPlaceHolder>" ; } isUsingDiskStore ( com . gemstone . gemfire . cache . Region , com . gemstone . gemfire . cache . DiskStore ) { return ( ( isPersistent ( region ) ) || ( isOverflowToDisk ( region ) ) ) && ( com . gemstone . gemfire . internal . lang . ObjectUtils . equals ( getDiskStoreName ( region ) , diskStore . getName ( ) ) ) ; }
org . junit . Assert . assertFalse ( function . isUsingDiskStore ( mockGateway , mockDiskStore ) )
testMappingFileStaticResourceLocationHasPrecedence ( ) { java . lang . String wrong = "some_file" ; configure ( staticResourceLocation , "some_location" ) ; configure ( resourceMappingFile , wrong ) ; java . lang . String mappingFile = org . richfaces . resource . ResourceMappingConfiguration . getResourceMappingFile ( ) ; "<AssertPlaceHolder>" ; } getResourceMappingFile ( ) { if ( ( org . richfaces . resource . ResourceMappingConfiguration . getStaticResourceLocation ( ) ) != null ) { return org . richfaces . resource . ResourceMappingConfiguration . DEFAULT_STATIC_RESOURCE_MAPPING_LOCATION ; } return org . richfaces . resource . ResourceMappingConfiguration . getConfiguration ( CoreConfiguration . Items . resourceMappingFile ) ; }
org . junit . Assert . assertEquals ( ResourceMappingConfiguration . DEFAULT_STATIC_RESOURCE_MAPPING_LOCATION , mappingFile )
serverMaxMessageSizeOkTest ( ) { java . util . Map < java . lang . String , java . lang . Object > serverProperties = getServerProperties ( ) ; serverProperties . put ( TyrusWebSocketEngine . INCOMING_BUFFER_SIZE , 3 ) ; final java . util . concurrent . atomic . AtomicBoolean warningLogged = new java . util . concurrent . atomic . AtomicBoolean ( false ) ; org . glassfish . tyrus . test . e2e . non_deployable . MaxMessageSizeDeploymentTest . LoggerHandler handler = new org . glassfish . tyrus . test . e2e . non_deployable . MaxMessageSizeDeploymentTest . LoggerHandler ( ) { @ org . glassfish . tyrus . test . e2e . non_deployable . Override public void publish ( java . util . logging . LogRecord record ) { System . out . println ( ( "Logged<sp>message:<sp>" + ( record . getMessage ( ) ) ) ) ; warningLogged . set ( true ) ; } } ; logger . setLevel ( Level . CONFIG ) ; logger . addHandler ( handler ) ; org . glassfish . tyrus . server . Server server = null ; try { server = startServer ( org . glassfish . tyrus . test . e2e . non_deployable . MaxMessageSizeDeploymentTest . LargeMaxMessageSizeServerEndpoint . class ) ; } finally { stopServer ( server ) ; } "<AssertPlaceHolder>" ; logger . removeHandler ( handler ) ; } get ( ) { if ( ( buffer ) == null ) { buffer = java . nio . ByteBuffer . allocate ( sslEngine . getSession ( ) . getPacketBufferSize ( ) ) ; } return buffer ; }
org . junit . Assert . assertFalse ( warningLogged . get ( ) )
testBigFilePutAndCancelJob ( ) { org . ourgrid . system . units . BrokerUnit brokerUnit = new org . ourgrid . system . units . BrokerUnit ( org . ourgrid . system . units . BrokerUnit . BROKER_PROPERTIES_FILENAME , 1 , 1 , 10 , 1 ) ; unitManager . addUnit ( brokerUnit ) ; org . ourgrid . system . units . WorkerUnit workerUnit = unitManager . buildNewUnit ( org . ourgrid . system . units . WorkerUnit . class ) ; workerUnit . setPlaypenRootPath ( org . ourgrid . system . TEMP_TEST_DIR ) ; workerUnit . setStorageRootPath ( org . ourgrid . system . TEMP_TEST_DIR ) ; org . ourgrid . system . units . PeerUnit peerUnit = unitManager . buildNewUnit ( org . ourgrid . system . units . PeerUnit . class ) ; brokerUnit . initKeys ( ) ; peerUnit . initKeys ( ) ; workerUnit . initKeys ( ) ; System . out . println ( "===><sp>Remote<sp>entities<sp>running<sp>(1<sp>worker,<sp>1<sp>peer,<sp>1<sp>broker)<sp><===" ) ; conditionExpecter . waitUntilConditionIsMet ( new org . ourgrid . system . condition . PeerHasTheWorkerInStateCondition ( peerUnit , workerUnit , org . ourgrid . common . interfaces . to . LocalWorkerState . IDLE ) ) ; System . out . println ( "===><sp>Set<sp>workers<sp><===" ) ; peerUnit . addUser ( brokerUnit . getLogin ( ) ) ; conditionExpecter . waitUntilConditionIsMet ( new org . ourgrid . system . condition . BrokerHasAPeerInTheState ( brokerUnit , peerUnit , PeerTestState . UP ) ) ; System . out . println ( "===><sp>Set<sp>peer<sp><===" ) ; java . io . File putSourceTempFile = org . ourgrid . common . util . TempFileManager . createTempFileWithBogusData ( getClass ( ) . getSimpleName ( ) , "" , tempFileDir , ( ( 1024 * 1024 ) * 5 ) ) ; java . lang . String [ ] putSources = new java . lang . String [ ] { putSourceTempFile . getPath ( ) } ; java . lang . String [ ] playpenDests = new java . lang . String [ ] { putSourceTempFile . getName ( ) } ; java . lang . String [ ] getDests = new java . lang . String [ ] { ( putSourceTempFile . getPath ( ) ) + "_get" } ; int jobID = brokerUnit . addJob ( org . ourgrid . system . units . UnitUtil . buildASmallSleepJobWithPuts ( putSources , playpenDests , getDests ) ) ; System . out . println ( "===><sp>job<sp>added<sp><===" ) ; conditionExpecter . waitUntilConditionIsMet ( new org . ourgrid . system . condition . BrokerJobRunningCondition ( brokerUnit , jobID ) ) ; java . lang . Thread . sleep ( 1000 ) ; brokerUnit . cancelJob ( jobID ) ; conditionExpecter . waitUntilConditionIsMet ( new org . ourgrid . system . condition . BrokerJobStateCondition ( brokerUnit , jobID , org . ourgrid . common . interfaces . to . GridProcessState . CANCELLED ) ) ; System . out . println ( "===><sp>job<sp>cancelled<sp><===" ) ; conditionExpecter . waitUntilConditionIsMet ( new org . ourgrid . system . condition . PeerHasTheWorkerInStateCondition ( peerUnit , workerUnit , org . ourgrid . common . interfaces . to . LocalWorkerState . IDLE ) ) ; System . out . println ( "===><sp>Worker<sp>is<sp>idle<sp><===" ) ; java . io . File getFile = new java . io . File ( getDests [ 0 ] ) ; "<AssertPlaceHolder>" ; } waitUntilConditionIsMet ( org . ourgrid . system . condition . Condition , long , int ) { int rounds = 0 ; while ( ! ( condition . isConditionMet ( ) ) ) { wait ( waitTime ) ; System . out . println ( ( ( ( "Testing:<sp>(" + ( condition . getClass ( ) . getSimpleName ( ) ) ) + ")" ) + ( condition . detailMessage ( ) ) ) ) ; if ( maxRounds == rounds ) { throw new java . lang . Exception ( ( "Condition<sp>not<sp>met<sp>in<sp>maximum<sp>number<sp>of<sp>rounds:<sp>" + ( condition . detailMessage ( ) ) ) ) ; } rounds ++ ; } System . out . println ( ( "Condition<sp>met:<sp>" + ( condition . getClass ( ) . getSimpleName ( ) ) ) ) ; }
org . junit . Assert . assertFalse ( getFile . exists ( ) )
testGetListOfJobDescription ( ) { er . quartzscheduler . util . ERQSSchedulerServiceFrameworkPrincipal fp = new er . quartzscheduler . util . ERQSSchedulerFP4Test ( ) ; "<AssertPlaceHolder>" ; } getListOfJobDescription ( com . webobjects . eocontrol . EOEditingContext ) { return com . webobjects . foundation . NSArray . emptyArray ( ) ; }
org . junit . Assert . assertTrue ( ( ( fp . getListOfJobDescription ( null ) . size ( ) ) == 0 ) )
testSerialization ( ) { org . jfree . data . xy . VectorDataItem v1 = new org . jfree . data . xy . VectorDataItem ( 1.0 , 2.0 , 3.0 , 4.0 ) ; org . jfree . data . xy . VectorDataItem v2 = ( ( org . jfree . data . xy . VectorDataItem ) ( org . jfree . chart . TestUtils . serialised ( v1 ) ) ) ; "<AssertPlaceHolder>" ; } serialised ( java . lang . Object ) { java . lang . Object result = null ; java . io . ByteArrayOutputStream buffer = new java . io . ByteArrayOutputStream ( ) ; java . io . ObjectOutput out ; try { out = new java . io . ObjectOutputStream ( buffer ) ; out . writeObject ( original ) ; out . close ( ) ; java . io . ObjectInput in = new java . io . ObjectInputStream ( new java . io . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; result = in . readObject ( ) ; in . close ( ) ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } catch ( java . lang . ClassNotFoundException e ) { throw new java . lang . RuntimeException ( e ) ; } return result ; }
org . junit . Assert . assertEquals ( v1 , v2 )
testSnifferPPTDirectGuess ( ) { java . io . File pptFile = org . nuxeo . ecm . platform . mimetype . detectors . TestDetector . getFileFromTestResource ( "test-data/hello.ppt" ) ; org . nuxeo . ecm . platform . mimetype . detectors . PptMimetypeSniffer sniffer = new org . nuxeo . ecm . platform . mimetype . detectors . PptMimetypeSniffer ( ) ; java . lang . String [ ] returnedMimetype = sniffer . guessPowerpoint ( pptFile ) ; "<AssertPlaceHolder>" ; } guessPowerpoint ( java . io . File ) { java . lang . String [ ] mimetypes = new java . lang . String [ ] { } ; try { try ( java . io . FileInputStream stream = new java . io . FileInputStream ( file ) ) { try ( org . apache . poi . hslf . usermodel . HSLFSlideShow ppt = new org . apache . poi . hslf . usermodel . HSLFSlideShow ( stream ) ) { if ( ( ppt . getSlides ( ) . size ( ) ) != 0 ) { mimetypes = getHandledTypes ( ) ; } } } } catch ( java . io . IOException | java . lang . RuntimeException e ) { org . nuxeo . ecm . platform . mimetype . detectors . PptMimetypeSniffer . log . debug ( "MimeType<sp>detector:<sp>Not<sp>a<sp>PowerPoint<sp>file" , e ) ; } return mimetypes ; }
org . junit . Assert . assertTrue ( ( ( returnedMimetype . length ) > 0 ) )
testScroll ( ) { openTestURL ( ) ; com . vaadin . testbench . TestBenchElement tab = ( ( com . vaadin . testbench . TestBenchElement ) ( getDriver ( ) . findElement ( org . openqa . selenium . By . className ( "v-tabsheet-tabitemcell" ) ) ) ) ; tab . click ( 10 , 10 ) ; org . openqa . selenium . Point oldLocation = tab . getLocation ( ) ; tab . sendKeys ( Keys . SPACE ) ; org . openqa . selenium . Point newLocation = tab . getLocation ( ) ; "<AssertPlaceHolder>" ; } sendKeys ( java . lang . CharSequence [ ] ) { sendKeys ( 50 , keysToSend ) ; }
org . junit . Assert . assertEquals ( oldLocation , newLocation )
should_count_with_no_attributes ( ) { org . apache . deltaspike . data . test . domain . Simple simple = testData . createSimple ( "testFindAll1" ) ; testData . createSimple ( "testFindAll2" ) ; javax . persistence . metamodel . SingularAttribute < org . apache . deltaspike . data . test . domain . Simple , java . lang . Object > [ ] attributes = new javax . persistence . metamodel . SingularAttribute [ ] { } ; java . lang . Long result = repo . count ( simple , attributes ) ; "<AssertPlaceHolder>" ; } valueOf ( java . lang . String ) { return org . apache . deltaspike . core . api . projectstage . ProjectStage . projectStages . get ( projectStageClassName ) ; }
org . junit . Assert . assertEquals ( java . lang . Long . valueOf ( 2 ) , result )
testWrapUnwrap ( ) { final javax . security . sasl . SaslClientFactory factory = obtainSaslClientFactory ( org . wildfly . security . sasl . external . ExternalSaslClientFactory . class ) ; final javax . security . sasl . SaslClient saslClient = factory . createSaslClient ( org . wildfly . security . sasl . external . ExternalSaslClientTest . MECHANISMS_EXTERNAL_ONLY , org . wildfly . security . sasl . external . ExternalSaslClientTest . ADMIN , "test" , "localhost" , setProps ( ) , null ) ; try { saslClient . wrap ( new byte [ ] { } , 0 , 0 ) ; org . junit . Assert . fail ( "wrap()<sp>invocation<sp>should<sp>throw<sp>IllegalStateException<sp>as<sp>not<sp>yet<sp>completed" ) ; } catch ( java . lang . IllegalStateException e ) { } try { saslClient . unwrap ( new byte [ ] { } , 0 , 0 ) ; org . junit . Assert . fail ( "unwrap()<sp>invocation<sp>should<sp>throw<sp>IllegalStateException<sp>as<sp>not<sp>yet<sp>completed" ) ; } catch ( java . lang . IllegalStateException e ) { } saslClient . evaluateChallenge ( org . wildfly . security . sasl . external . ExternalSaslClientTest . BYTES_EMPTY ) ; "<AssertPlaceHolder>" ; try { saslClient . wrap ( new byte [ ] { } , 0 , 0 ) ; org . junit . Assert . fail ( "wrap()<sp>invocation<sp>should<sp>throw<sp>IllegalStateException<sp>as<sp>this<sp>mechanism<sp>supports<sp>neither<sp>integrity<sp>nor<sp>privacy" ) ; } catch ( java . lang . IllegalStateException e ) { } try { saslClient . unwrap ( new byte [ ] { } , 0 , 0 ) ; org . junit . Assert . fail ( "wrap()<sp>invocation<sp>should<sp>throw<sp>IllegalStateException<sp>as<sp>this<sp>mechanism<sp>supports<sp>neither<sp>integrity<sp>nor<sp>privacy" ) ; } catch ( java . lang . IllegalStateException e ) { } } isComplete ( ) { return delegate . isComplete ( ) ; }
org . junit . Assert . assertTrue ( saslClient . isComplete ( ) )
givenBigIntegerNumbers_whenAddedTogether_thenGetExpectedResult ( ) { java . math . BigInteger numStarsMilkyWay = new java . math . BigInteger ( "8731409320171337804361260816606476" ) ; java . math . BigInteger numStarsAndromeda = new java . math . BigInteger ( "5379309320171337804361260816606476" ) ; java . math . BigInteger totalStars = numStarsMilkyWay . add ( numStarsAndromeda ) ; java . math . BigInteger result = new java . math . BigInteger ( "14110718640342675608722521633212952" ) ; "<AssertPlaceHolder>" ; } add ( org . springframework . boot . actuate . audit . AuditEvent ) { if ( ( ! ( com . baeldung . jhipster5 . repository . CustomAuditEventRepository . AUTHORIZATION_FAILURE . equals ( event . getType ( ) ) ) ) && ( ! ( Constants . ANONYMOUS_USER . equals ( event . getPrincipal ( ) ) ) ) ) { com . baeldung . jhipster5 . domain . PersistentAuditEvent persistentAuditEvent = new com . baeldung . jhipster5 . domain . PersistentAuditEvent ( ) ; persistentAuditEvent . setPrincipal ( event . getPrincipal ( ) ) ; persistentAuditEvent . setAuditEventType ( event . getType ( ) ) ; persistentAuditEvent . setAuditEventDate ( event . getTimestamp ( ) ) ; com . baeldung . jhipster5 . repository . Map < java . lang . String , java . lang . String > eventData = auditEventConverter . convertDataToStrings ( event . getData ( ) ) ; persistentAuditEvent . setData ( truncate ( eventData ) ) ; persistenceAuditEventRepository . save ( persistentAuditEvent ) ; } }
org . junit . Assert . assertEquals ( result , totalStars )
testEmptyPortEquivalentToDefaultPortForHttpsAbsoluteURI ( ) { final org . apache . hc . core5 . http . HttpHost host = new org . apache . hc . core5 . http . HttpHost ( "https" , "foo.example.com" , ( - 1 ) ) ; final org . apache . hc . client5 . http . classic . methods . HttpGet get1 = new org . apache . hc . client5 . http . classic . methods . HttpGet ( "https://bar.example.com:/" ) ; final org . apache . hc . client5 . http . classic . methods . HttpGet get2 = new org . apache . hc . client5 . http . classic . methods . HttpGet ( "https://bar.example.com:443/" ) ; final java . lang . String uri1 = extractor . generateKey ( host , get1 ) ; final java . lang . String uri2 = extractor . generateKey ( host , get2 ) ; "<AssertPlaceHolder>" ; } generateKey ( org . apache . hc . core5 . http . HttpHost , org . apache . hc . core5 . http . HttpRequest ) { final java . lang . String s = org . apache . hc . client5 . http . impl . cache . HttpCacheSupport . getRequestUri ( request , host ) ; try { return generateKey ( new java . net . URI ( s ) ) ; } catch ( final java . net . URISyntaxException ex ) { return s ; } }
org . junit . Assert . assertEquals ( uri1 , uri2 )
conversionUnsignedInteger32 ( ) { jrds . snmp . jrds . snmp . SnmpVars vars = new jrds . snmp . jrds . snmp . SnmpVars ( ) ; org . snmp4j . smi . OID oid1 = new org . snmp4j . smi . OID ( "1" ) ; org . snmp4j . smi . VariableBinding vb = new org . snmp4j . smi . VariableBinding ( oid1 , new org . snmp4j . smi . UnsignedInteger32 ( 1 ) ) ; vars . addVariable ( vb ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { javax . xml . xpath . XPathExpression e = jrds . factories . xml . CompiledXPath . xpc . get ( xpath ) ; if ( e == null ) { jrds . factories . xml . CompiledXPath . logger . debug ( jrds . Util . delayedFormatString ( "Uncompiled<sp>xpath:<sp>%s" , xpath ) ) ; e = jrds . factories . xml . CompiledXPath . xpather . get ( ) . compile ( xpath ) ; jrds . factories . xml . CompiledXPath . xpc . put ( xpath , e ) ; } return e ; }
org . junit . Assert . assertEquals ( vars . get ( oid1 ) , ( ( long ) ( 1 ) ) )
testGraphMerge ( ) { java . util . List < java . lang . String > expectedOrder = new java . util . ArrayList ( ) ; expectedOrder . add ( "A" ) ; expectedOrder . add ( "I" ) ; expectedOrder . add ( "B" ) ; expectedOrder . add ( "C" ) ; expectedOrder . add ( "H" ) ; expectedOrder . add ( "D" ) ; expectedOrder . add ( "G" ) ; expectedOrder . add ( "E" ) ; expectedOrder . add ( "F" ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphA = createGraph ( "A" ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphI = createGraph ( "I" ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphB = createGraph ( "B" ) ; graphA . addDependentGraph ( graphB ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphC = createGraph ( "C" ) ; graphA . addDependentGraph ( graphC ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphH = createGraph ( "H" ) ; graphI . addDependentGraph ( graphH ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphG = createGraph ( "G" ) ; graphC . addDependentGraph ( graphG ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphE = createGraph ( "E" ) ; graphB . addDependentGraph ( graphE ) ; graphG . addDependentGraph ( graphE ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphD = createGraph ( "D" ) ; graphB . addDependentGraph ( graphD ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > graphF = createGraph ( "F" ) ; graphD . addDependentGraph ( graphF ) ; graphE . addDependentGraph ( graphF ) ; graphH . addDependentGraph ( graphF ) ; com . microsoft . azure . management . resources . fluentcore . dag . DAGraph < java . lang . String , com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder > dag = graphF ; dag . prepareForEnumeration ( ) ; com . microsoft . azure . management . resources . fluentcore . dag . ItemHolder nextNode = dag . getNext ( ) ; int i = 0 ; while ( nextNode != null ) { "<AssertPlaceHolder>" ; dag . reportCompletion ( nextNode ) ; nextNode = dag . getNext ( ) ; i ++ ; } } get ( java . lang . String ) { return this . getAsync ( webhookName ) . toBlocking ( ) . single ( ) ; }
org . junit . Assert . assertEquals ( expectedOrder . get ( i ) , nextNode . key ( ) )
multipleReaders ( ) { final int numberOfThreads = org . exist . storage . lock . CollectionLocksTest . CONCURRENCY_LEVEL ; final org . exist . xmldb . XmldbURI collectionUri = org . exist . xmldb . XmldbURI . create ( "/db/x/y/z" ) ; final org . exist . storage . lock . LockManager lockManager = new org . exist . storage . lock . LockManager ( org . exist . storage . lock . CollectionLocksTest . CONCURRENCY_LEVEL ) ; final org . exist . storage . lock . CountDownLatch continueLatch = new org . exist . storage . lock . CountDownLatch ( numberOfThreads ) ; final java . util . function . Supplier < org . exist . storage . lock . Callable < java . lang . Void > > readCollectionFn = ( ) -> ( ) -> { try ( final org . exist . storage . lock . ManagedCollectionLock collectionLock = lockManager . acquireCollectionReadLock ( collectionUri ) ) { continueLatch . countDown ( ) ; continueLatch . await ( ) ; } return null ; } ; final java . util . List < org . exist . storage . lock . Callable < java . lang . Void > > callables = new java . util . ArrayList ( ) ; for ( int i = 0 ; i < numberOfThreads ; i ++ ) { callables . add ( readCollectionFn . get ( ) ) ; } final org . exist . storage . lock . ExecutorService executorService = org . exist . storage . lock . Executors . newFixedThreadPool ( numberOfThreads ) ; final java . util . List < org . exist . storage . lock . Future < java . lang . Void > > futures = executorService . invokeAll ( callables , org . exist . storage . lock . CollectionLocksTest . MULTI_READER_TEST_TIMEOUT , TimeUnit . MILLISECONDS ) ; for ( final org . exist . storage . lock . Future < java . lang . Void > future : futures ) { if ( future . isCancelled ( ) ) { org . junit . Assert . fail ( "multipleReaders<sp>test<sp>likely<sp>showed<sp>a<sp>thread<sp>deadlock" ) ; } } executorService . shutdown ( ) ; "<AssertPlaceHolder>" ; } getCount ( ) { return count ; }
org . junit . Assert . assertEquals ( 0 , continueLatch . getCount ( ) )
testIsWhitespaceXml ( ) { final java . io . File xml = new java . io . File ( getClass ( ) . getResource ( "/whitespace-followed-by-xml.jtl" ) . toURI ( ) ) ; final boolean result = hudson . plugins . performance . parsers . JMeterParser . isXmlFile ( xml ) ; "<AssertPlaceHolder>" ; } isXmlFile ( java . io . File ) { try ( java . io . FileReader fr = new java . io . FileReader ( file ) ; java . io . BufferedReader reader = new java . io . BufferedReader ( fr ) ) { java . lang . String line ; boolean isXml = false ; while ( ( line = reader . readLine ( ) ) != null ) { if ( ( line . trim ( ) . length ( ) ) == 0 ) { continue ; } if ( line . toLowerCase ( ) . trim ( ) . startsWith ( "<?xml<sp>" ) ) { isXml = true ; } break ; } return isXml ; } }
org . junit . Assert . assertTrue ( result )
entityTwoPrimNoMetadata ( ) { final org . apache . olingo . commons . api . edm . EdmEntitySet edmEntitySet = org . apache . olingo . server . core . serializer . json . ODataJsonSerializerTest . entityContainer . getEntitySet ( "ESTwoPrim" ) ; final org . apache . olingo . commons . api . data . Entity entity = data . readAll ( edmEntitySet ) . getEntities ( ) . get ( 0 ) ; final java . lang . String resultString = org . apache . commons . io . IOUtils . toString ( serializerNoMetadata . entity ( org . apache . olingo . server . core . serializer . json . ODataJsonSerializerTest . metadata , edmEntitySet . getEntityType ( ) , entity , null ) . getContent ( ) ) ; final java . lang . String expectedResult = "{\"PropertyInt16\":32766,\"PropertyString\":\"Test<sp>String1\"}" ; "<AssertPlaceHolder>" ; } getContent ( ) { return new java . io . ByteArrayInputStream ( content . toByteArray ( ) ) ; }
org . junit . Assert . assertEquals ( expectedResult , resultString )
testGetDeviceByIDValid ( ) { network . addRemoteDevice ( remoteDevice1 ) ; network . addRemoteDevice ( remoteDevice3 ) ; com . digi . xbee . api . RemoteXBeeDevice found = network . getDevice ( com . digi . xbee . api . XBeeNetworkGetTest . NODE_ID ) ; "<AssertPlaceHolder>" ; } getDevice ( com . digi . xbee . api . models . XBee16BitAddress ) { if ( ( localDevice . getXBeeProtocol ( ) ) == ( com . digi . xbee . api . models . XBeeProtocol . DIGI_MESH ) ) throw new com . digi . xbee . api . exceptions . OperationNotSupportedException ( "DigiMesh<sp>protocol<sp>does<sp>not<sp>support<sp>16-bit<sp>addressing." ) ; if ( ( localDevice . getXBeeProtocol ( ) ) == ( com . digi . xbee . api . models . XBeeProtocol . DIGI_POINT ) ) throw new com . digi . xbee . api . exceptions . OperationNotSupportedException ( "Point-to-Multipoint<sp>protocol<sp>does<sp>not<sp>support<sp>16-bit<sp>addressing." ) ; if ( address == null ) throw new java . lang . NullPointerException ( "16-bit<sp>address<sp>cannot<sp>be<sp>null." ) ; if ( address . equals ( XBee16BitAddress . UNKNOWN_ADDRESS ) ) throw new java . lang . IllegalArgumentException ( "16-bit<sp>address<sp>cannot<sp>be<sp>unknown." ) ; logger . debug ( "{}Getting<sp>device<sp>'{}'<sp>from<sp>network." , localDevice . toString ( ) , address ) ; com . digi . xbee . api . RemoteXBeeDevice devInNetwork = null ; java . util . Collection < com . digi . xbee . api . RemoteXBeeDevice > devices = remotesBy64BitAddr . values ( ) ; for ( com . digi . xbee . api . RemoteXBeeDevice d : devices ) { com . digi . xbee . api . models . XBee16BitAddress a = get16BitAddress ( d ) ; if ( ( a != null ) && ( a . equals ( address ) ) ) { devInNetwork = d ; break ; } } if ( devInNetwork == null ) devInNetwork = remotesBy16BitAddr . get ( address ) ; return devInNetwork ; }
org . junit . Assert . assertEquals ( remoteDevice1 , found )
testParseBundleWithBinary ( ) { ca . uhn . fhir . parser . Binary patient = new ca . uhn . fhir . parser . Binary ( ) ; patient . setId ( new ca . uhn . fhir . parser . IdType ( "http://base/Binary/11/_history/22" ) ) ; patient . setContentType ( "foo" ) ; patient . setContent ( new byte [ ] { 1 , 2 , 3 , 4 } ) ; java . lang . String val = ca . uhn . fhir . parser . JsonParserDstu3Test . ourCtx . newJsonParser ( ) . encodeResourceToString ( patient ) ; java . lang . String expected = "{\"resourceType\":\"Binary\",\"id\":\"11\",\"meta\":{\"versionId\":\"22\"},\"contentType\":\"foo\",\"content\":\"AQIDBA==\"}" ; ca . uhn . fhir . parser . JsonParserDstu3Test . ourLog . info ( "Expected:<sp>{}" , expected ) ; ca . uhn . fhir . parser . JsonParserDstu3Test . ourLog . info ( "Actual<sp>:<sp>{}" , val ) ; "<AssertPlaceHolder>" ; } newJsonParser ( ) { return new ca . uhn . fhir . context . JsonParser ( this , myParserErrorHandler ) ; }
org . junit . Assert . assertEquals ( expected , val )
testDistributedCacheWithIterations ( ) { final java . lang . String testString = "Et<sp>tu,<sp>Brute?" ; final java . lang . String testName = "testing_caesar" ; final java . io . File folder = org . apache . flink . test . iterative . aggregators . AggregatorsITCase . tempFolder . newFolder ( ) ; final java . io . File resultFile = new java . io . File ( folder , java . util . UUID . randomUUID ( ) . toString ( ) ) ; java . lang . String testPath = resultFile . toString ( ) ; java . lang . String resultPath = resultFile . toURI ( ) . toString ( ) ; java . io . File tempFile = new java . io . File ( testPath ) ; try ( java . io . FileWriter writer = new java . io . FileWriter ( tempFile ) ) { writer . write ( testString ) ; } final org . apache . flink . api . java . ExecutionEnvironment env = org . apache . flink . api . java . ExecutionEnvironment . getExecutionEnvironment ( ) ; env . registerCachedFile ( resultPath , testName ) ; org . apache . flink . api . java . operators . IterativeDataSet < java . lang . Long > solution = env . fromElements ( 1L ) . iterate ( 2 ) ; solution . closeWith ( env . generateSequence ( 1 , 2 ) . filter ( new org . apache . flink . api . common . functions . RichFilterFunction < java . lang . Long > ( ) { @ org . apache . flink . test . iterative . aggregators . Override public void open ( org . apache . flink . configuration . Configuration parameters ) throws java . lang . Exception { java . io . File file = getRuntimeContext ( ) . getDistributedCache ( ) . getFile ( testName ) ; java . io . BufferedReader reader = new java . io . BufferedReader ( new java . io . FileReader ( file ) ) ; java . lang . String output = reader . readLine ( ) ; reader . close ( ) ; "<AssertPlaceHolder>" ; } @ org . apache . flink . test . iterative . aggregators . Override public boolean filter ( java . lang . Long value ) throws java . lang . Exception { return false ; } } ) . withBroadcastSet ( solution , "SOLUTION" ) ) . output ( new org . apache . flink . api . java . io . DiscardingOutputFormat < java . lang . Long > ( ) ) ; env . execute ( ) ; } close ( ) { if ( ( uploadDirectory ) != null ) { org . apache . flink . util . FileUtils . deleteDirectory ( uploadDirectory . toFile ( ) ) ; } }
org . junit . Assert . assertEquals ( output , testString )
testAddKVColumnToExistingFamily ( ) { long ts = nextTimestamp ( ) ; java . lang . String tenantId = getOrganizationId ( ) ; initATableValues ( tenantId , getDefaultSplits ( tenantId ) , null , ts ) ; java . util . Properties props = new java . util . Properties ( ) ; props . setProperty ( PhoenixRuntime . CURRENT_SCN_ATTRIB , java . lang . Long . toString ( ( ts + 5 ) ) ) ; java . sql . Connection conn1 = java . sql . DriverManager . getConnection ( com . salesforce . phoenix . end2end . PHOENIX_JDBC_URL , props ) ; conn1 . createStatement ( ) . executeUpdate ( ( ( "ALTER<sp>TABLE<sp>" + ( ATABLE_NAME ) ) + "<sp>ADD<sp>z_integer<sp>integer" ) ) ; conn1 . close ( ) ; props . setProperty ( PhoenixRuntime . CURRENT_SCN_ATTRIB , java . lang . Long . toString ( ( ts + 6 ) ) ) ; java . sql . Connection conn2 = java . sql . DriverManager . getConnection ( com . salesforce . phoenix . end2end . PHOENIX_JDBC_URL , props ) ; java . lang . String query = "SELECT<sp>z_integer<sp>FROM<sp>aTable" ; "<AssertPlaceHolder>" ; conn2 . close ( ) ; props . setProperty ( PhoenixRuntime . CURRENT_SCN_ATTRIB , java . lang . Long . toString ( ( ts + 3 ) ) ) ; java . sql . Connection conn3 = java . sql . DriverManager . getConnection ( com . salesforce . phoenix . end2end . PHOENIX_JDBC_URL , props ) ; try { conn3 . prepareStatement ( query ) . executeQuery ( ) . next ( ) ; org . junit . Assert . fail ( ) ; } catch ( com . salesforce . phoenix . schema . ColumnNotFoundException e ) { } } prepareStatement ( java . lang . String ) { com . salesforce . phoenix . jdbc . PhoenixPreparedStatement statement = new com . salesforce . phoenix . jdbc . PhoenixPreparedStatement ( this , sql ) ; statements . add ( statement ) ; return statement ; }
org . junit . Assert . assertTrue ( conn2 . prepareStatement ( query ) . executeQuery ( ) . next ( ) )
AccessInstanceVariableSuperSuperClass ( ) { java . lang . String fromClass = "domain.direct.violating.AccessInstanceVariableSuperSuperClass" ; java . lang . String toClass = "technology.direct.subclass.CallInstanceSubSubClassDOA" ; java . util . ArrayList < java . lang . String > typesToFind = new java . util . ArrayList < java . lang . String > ( ) ; typesToFind . add ( "Access" ) ; boolean outcome1 = areDependencyTypesDetected ( fromClass , toClass , typesToFind , "Instance<sp>Variable" , true ) ; toClass = "technology.direct.subclass.CallInstanceSubClassDAO" ; boolean outcome2 = areDependencyTypesDetected ( fromClass , toClass , typesToFind , "Instance<sp>Variable" , true ) ; toClass = "technology.direct.dao.CallInstanceSuperClassDAO" ; boolean outcome3 = areDependencyTypesDetected ( fromClass , toClass , typesToFind , "Instance<sp>Variable" , true ) ; boolean totalOutcome = false ; if ( ( ( outcome1 == true ) || ( outcome2 == true ) ) || ( outcome3 == true ) ) { totalOutcome = true ; } "<AssertPlaceHolder>" ; } areDependencyTypesDetected ( java . lang . String , java . lang . String , java . util . ArrayList , java . lang . String , boolean ) { boolean dependencyTypesDetected = false ; java . util . TreeMap < java . lang . String , java . lang . Boolean > foundDependencyTypes = new java . util . TreeMap < java . lang . String , java . lang . Boolean > ( ) ; husaccttest . analyse . Java_AccuracyTestDependencyDetection . analyseService = husacct . ServiceProvider . getInstance ( ) . getAnalyseService ( ) ; husacct . common . dto . DependencyDTO [ ] foundDependencies = husaccttest . analyse . Java_AccuracyTestDependencyDetection . analyseService . getDependenciesFromClassToClass ( classFrom , classTo ) ; int numberOfDependencies = foundDependencies . length ; for ( java . lang . String dependencyType : dependencyTypes ) { boolean found = false ; for ( int i = 0 ; i < numberOfDependencies ; i ++ ) { if ( ( foundDependencies [ i ] . type . equals ( dependencyType ) ) && ( ( foundDependencies [ i ] . isIndirect ) == isIndirect ) ) { if ( ! ( subType . equals ( "" ) ) ) { if ( foundDependencies [ i ] . subType . equals ( subType ) ) { found = true ; } } else { found = true ; } } } foundDependencyTypes . put ( dependencyType , found ) ; } if ( ! ( foundDependencyTypes . containsValue ( false ) ) ) { dependencyTypesDetected = true ; } return dependencyTypesDetected ; }
org . junit . Assert . assertTrue ( totalOutcome )
testGetErrorInteger ( ) { lineData . put ( "key" , "value" ) ; "<AssertPlaceHolder>" ; } getInt ( java . lang . String ) { java . lang . String value = map . get ( key ) ; if ( value == null ) { return 0 ; } try { return java . lang . Integer . valueOf ( value ) ; } catch ( java . lang . NumberFormatException e ) { return 0 ; } }
org . junit . Assert . assertEquals ( 0 , lineData . getInt ( "key" ) )
alreadyRegistered_SameHostId ( ) { final boolean success = registrar . tryToRegister ( client ) ; "<AssertPlaceHolder>" ; verify ( client , never ( ) ) . createAndSetData ( idPath , hostId . getBytes ( ) ) ; } tryToRegister ( com . spotify . helios . servicescommon . coordination . ZooKeeperClient ) { client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . configHosts ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . configJobs ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . configJobRefs ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . statusHosts ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . statusMasters ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . historyJobs ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . configDeploymentGroups ( ) ) ; client . ensurePath ( com . spotify . helios . servicescommon . coordination . Paths . statusDeploymentGroups ( ) ) ; if ( ( upNode ) == null ) { final java . lang . String upPath = com . spotify . helios . servicescommon . coordination . Paths . statusMasterUp ( name ) ; client . ensurePath ( upPath , true ) ; upNode = client . persistentEphemeralNode ( upPath , Mode . EPHEMERAL , new byte [ ] { } ) ; upNode . start ( ) ; } com . spotify . helios . master . MasterZooKeeperRegistrar . log . info ( "ZooKeeper<sp>registration<sp>complete" ) ; return true ; }
org . junit . Assert . assertTrue ( success )