input stringlengths 28 18.7k | output stringlengths 39 1.69k |
|---|---|
copyInputStreamValidOutputStreamPosBufSz ( ) { java . io . ByteArrayOutputStream outputStream = new org . apache . maven . shared . utils . io . IOUtilTest . DontCloseByteArrayOutputStream ( ) ; byte [ ] input = new byte [ ] { 1 , 2 , 3 , 4 , 5 , 6 } ; org . apache . maven . shared . utils . io . IOUtil . copy ( new org . apache . maven . shared . utils . io . IOUtilTest . DontCloseByteArrayInputStream ( input ) , outputStream , 1 ) ; "<AssertPlaceHolder>" ; } copy ( java . io . InputStream , java . io . OutputStream , int ) { final byte [ ] buffer = new byte [ bufferSize ] ; int n ; while ( ( - 1 ) != ( n = input . read ( buffer ) ) ) { output . write ( buffer , 0 , n ) ; } } | org . junit . Assert . assertThat ( outputStream . toByteArray ( ) , org . hamcrest . CoreMatchers . is ( input ) ) |
testReplicationWhenBlockCorruption ( ) { org . apache . hadoop . hdfs . MiniDFSCluster cluster = null ; try { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . HdfsConfiguration ( ) ; conf . setLong ( DFSConfigKeys . DFS_NAMENODE_RECONSTRUCTION_PENDING_TIMEOUT_SEC_KEY , 1 ) ; cluster = new org . apache . hadoop . hdfs . MiniDFSCluster . Builder ( conf ) . numDataNodes ( 3 ) . storagesPerDatanode ( 1 ) . build ( ) ; org . apache . hadoop . fs . FileSystem fs = cluster . getFileSystem ( ) ; org . apache . hadoop . fs . Path filePath = new org . apache . hadoop . fs . Path ( "/test" ) ; org . apache . hadoop . fs . FSDataOutputStream create = fs . create ( filePath ) ; fs . setReplication ( filePath , ( ( short ) ( 1 ) ) ) ; create . write ( new byte [ 1024 ] ) ; create . close ( ) ; org . apache . hadoop . hdfs . protocol . ExtendedBlock block = org . apache . hadoop . hdfs . DFSTestUtil . getFirstBlock ( fs , filePath ) ; int numReplicaCreated = 0 ; for ( final org . apache . hadoop . hdfs . server . datanode . DataNode dn : cluster . getDataNodes ( ) ) { if ( ! ( dn . getFSDataset ( ) . contains ( block ) ) ) { cluster . getFsDatasetTestUtils ( dn ) . injectCorruptReplica ( block ) ; numReplicaCreated ++ ; } } "<AssertPlaceHolder>" ; fs . setReplication ( filePath , ( ( short ) ( 3 ) ) ) ; cluster . restartDataNodes ( ) ; cluster . waitActive ( ) ; cluster . triggerBlockReports ( ) ; org . apache . hadoop . hdfs . DFSTestUtil . waitReplication ( fs , filePath , ( ( short ) ( 3 ) ) ) ; } finally { if ( cluster != null ) { cluster . shutdown ( ) ; } } } injectCorruptReplica ( org . apache . hadoop . hdfs . protocol . ExtendedBlock ) { com . google . common . base . Preconditions . checkState ( ( ! ( dataset . contains ( block ) ) ) , ( ( "Block<sp>" + block ) + "<sp>already<sp>exists<sp>on<sp>dataset." ) ) ; try ( org . apache . hadoop . hdfs . server . datanode . fsdataset . FsDatasetSpi . FsVolumeReferences volRef = dataset . getFsVolumeReferences ( ) ) { org . apache . hadoop . hdfs . server . datanode . fsdataset . impl . FsVolumeImpl volume = ( ( org . apache . hadoop . hdfs . server . datanode . fsdataset . impl . FsVolumeImpl ) ( volRef . get ( 0 ) ) ) ; org . apache . hadoop . hdfs . server . datanode . FinalizedReplica finalized = new org . apache . hadoop . hdfs . server . datanode . FinalizedReplica ( block . getLocalBlock ( ) , volume , volume . getFinalizedDir ( block . getBlockPoolId ( ) ) ) ; java . io . File blockFile = finalized . getBlockFile ( ) ; if ( ! ( blockFile . createNewFile ( ) ) ) { throw new org . apache . commons . io . FileExistsException ( ( ( "Block<sp>file<sp>" + blockFile ) + "<sp>already<sp>exists." ) ) ; } java . io . File metaFile = org . apache . hadoop . hdfs . server . datanode . fsdataset . impl . FsDatasetUtil . getMetaFile ( blockFile , 1000 ) ; if ( ! ( metaFile . createNewFile ( ) ) ) { throw new org . apache . commons . io . FileExistsException ( ( ( "Meta<sp>file<sp>" + metaFile ) + "<sp>already<sp>exists." ) ) ; } dataset . volumeMap . add ( block . getBlockPoolId ( ) , finalized ) ; } } | org . junit . Assert . assertEquals ( 2 , numReplicaCreated ) |
createEnrollmentGroupQueryWithPageSizeSucceed ( tests . unit . com . microsoft . azure . sdk . iot . provisioning . service . QuerySpecification ) { tests . unit . com . microsoft . azure . sdk . iot . provisioning . service . ProvisioningServiceClient provisioningServiceClient = createClient ( ) ; new mockit . NonStrictExpectations ( ) { { mockit . Deencapsulation . invoke ( mockedEnrollmentGroupManager , "createQuery" , mockedQuerySpecification , 10 ) ; times = 1 ; } } ; tests . unit . com . microsoft . azure . sdk . iot . provisioning . service . Query query = provisioningServiceClient . createEnrollmentGroupQuery ( mockedQuerySpecification , 10 ) ; "<AssertPlaceHolder>" ; } createEnrollmentGroupQuery ( com . microsoft . azure . sdk . iot . provisioning . service . QuerySpecification , int ) { return enrollmentGroupManager . createQuery ( querySpecification , pageSize ) ; } | org . junit . Assert . assertNotNull ( query ) |
testGradient ( ) { cn . edu . hit . ir . JNN . Model m = new cn . edu . hit . ir . JNN . Model ( ) ; cn . edu . hit . ir . JNN . Trainers . SimpleSGDTrainer sgd = new cn . edu . hit . ir . JNN . Trainers . SimpleSGDTrainer ( m ) ; cn . edu . hit . ir . JNN . ComputationGraph cg = new cn . edu . hit . ir . JNN . ComputationGraph ( ) ; java . util . Vector < java . lang . Double > xV = new java . util . Vector < java . lang . Double > ( java . util . Arrays . asList ( 1.0 , 2.0 ) ) ; java . util . Vector < java . lang . Double > yV = new java . util . Vector < java . lang . Double > ( java . util . Arrays . asList ( 1.0 , 2.0 ) ) ; cn . edu . hit . ir . JNN . Expression W = Expression . Creator . parameter ( cg , m . addParameters ( cn . edu . hit . ir . JNN . Dim . create ( 2 , 1 ) ) ) ; cn . edu . hit . ir . JNN . Expression x = Expression . Creator . input ( cg , cn . edu . hit . ir . JNN . Dim . create ( 2 , 1 ) , xV ) ; cn . edu . hit . ir . JNN . Expression y = Expression . Creator . input ( cg , cn . edu . hit . ir . JNN . Dim . create ( 2 , 1 ) , yV ) ; cn . edu . hit . ir . JNN . Expression CM = Expression . Creator . cwiseMultiply ( x , W ) ; cn . edu . hit . ir . JNN . Expression loss = Expression . Creator . squaredDistance ( CM , y ) ; cg . gradientCheck ( ) ; cg . forward ( ) ; cg . backward ( ) ; "<AssertPlaceHolder>" ; } gradientCheck ( ) { for ( int i = 0 ; i < ( parameterNodes . size ( ) ) ; ++ i ) { int j = parameterNodes . get ( i ) ; if ( ( nodes . get ( j ) . getName ( ) ) == "ParameterNode" ) { checkParameterNode ( nodes . get ( j ) ) ; } else { checkLookupNode ( nodes . get ( j ) ) ; } } } | org . junit . Assert . assertEquals ( true , m . gradientCheck ( ) ) |
testGetDefinitions ( ) { org . eclipse . tracecompass . ctf . core . event . types . IDefinition result = fixture . getCurrentField ( ) ; "<AssertPlaceHolder>" ; } getCurrentField ( ) { return fDefinition ; } | org . junit . Assert . assertNotNull ( result ) |
doubleQuote ( ) { try { final java . lang . String content = this . parser . getTextFromFile ( com . github . jknack . antlr4ide . issues . Issue42 . class ) ; final org . eclipse . xtext . parser . IParseResult parseResults = this . parser . parse ( content ) ; final boolean syntaxErrors = parseResults . hasSyntaxErrors ( ) ; "<AssertPlaceHolder>" ; } catch ( final java . lang . Throwable throwable ) { throw org . eclipse . xtext . xbase . lib . Exceptions . sneakyThrow ( throwable ) ; } } parse ( java . lang . CharSequence ) { final java . lang . String content = input . toString ( ) ; final java . io . StringReader stringReader = new java . io . StringReader ( content ) ; return this . parser . parse ( stringReader ) ; } | org . junit . Assert . assertFalse ( syntaxErrors ) |
checkMessageMethod ( ) { java . lang . String str = "test" ; info . smart_tools . smartactors . feature_loading_system . interfaces . iplugin_loader . exception . PluginLoaderException exception = new info . smart_tools . smartactors . feature_loading_system . interfaces . iplugin_loader . exception . PluginLoaderException ( str ) ; "<AssertPlaceHolder>" ; throw exception ; } getMessage ( ) { return message ; } | org . junit . Assert . assertEquals ( exception . getMessage ( ) , str ) |
test2 ( ) { final java . util . List < io . vertigo . dynamo . plugins . environment . dsl . entity . DslEntity > entities = dslDefinitionRepository . getGrammar ( ) . getEntities ( ) ; final io . vertigo . dynamo . plugins . environment . dsl . entity . DslEntity entity = io . vertigo . dynamo . environment . plugins . loaders . kpr . definition . DslDefinitionBodyRuleTest . find ( entities , "Domain" ) ; final io . vertigo . dynamo . plugins . environment . loaders . kpr . definition . DslDefinitionBody definitionBody = new io . vertigo . dynamo . plugins . environment . loaders . kpr . rules . DslDefinitionBodyRule ( entity ) . parse ( "{<sp>dataType<sp>:<sp>String<sp>,<sp>formatter<sp>:<sp>FMT_DEFAULT,<sp>constraint<sp>:<sp>[<sp>CK_CODE_POSTAL<sp>]<sp>}<sp>" , 0 ) . getValue ( ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return value ; } | org . junit . Assert . assertNotNull ( definitionBody ) |
getPatient_shouldReturnNullWhenPersonDoesNotExist ( ) { executeDataSet ( org . openmrs . api . PatientServiceTest . USER_WHO_IS_NOT_PATIENT_XML ) ; org . openmrs . Patient patient = org . openmrs . api . PatientServiceTest . patientService . getPatientOrPromotePerson ( ( - 1 ) ) ; "<AssertPlaceHolder>" ; } getPatientOrPromotePerson ( java . lang . Integer ) { org . openmrs . Patient patient = null ; try { patient = org . openmrs . api . context . Context . getPatientService ( ) . getPatient ( patientOrPersonId ) ; } catch ( java . lang . ClassCastException ex ) { } if ( patient == null ) { org . openmrs . Person toPromote = org . openmrs . api . context . Context . getPersonService ( ) . getPerson ( patientOrPersonId ) ; if ( toPromote != null ) { patient = new org . openmrs . Patient ( toPromote ) ; } } return patient ; } | org . junit . Assert . assertNull ( patient ) |
testFilterLess ( ) { io . katharsis . queryspec . QuerySpec querySpec = new io . katharsis . queryspec . QuerySpec ( io . katharsis . jpa . model . TestEntity . class ) ; querySpec . addFilter ( new io . katharsis . queryspec . FilterSpec ( java . util . Arrays . asList ( "longValue" ) , io . katharsis . queryspec . FilterOperator . LT , 2 ) ) ; java . util . List < io . katharsis . jpa . model . TestEntity > list = repo . findAll ( querySpec ) ; "<AssertPlaceHolder>" ; } size ( ) { return list . size ( ) ; } | org . junit . Assert . assertEquals ( 2 , list . size ( ) ) |
testNoRequiredDocuments ( ) { try { getSimpleController ( initAttributes ) . process ( processingAttributes , getComponentClass ( ) ) ; org . junit . Assert . fail ( "Should<sp>fail<sp>with<sp>an<sp>exception." ) ; } catch ( org . carrot2 . core . ProcessingException e ) { "<AssertPlaceHolder>" . contains ( "No<sp>value<sp>for<sp>required<sp>attribute" ) ; } } getComponentClass ( ) { return org . carrot2 . clustering . stc . STCClusteringAlgorithm . class ; } | org . junit . Assert . assertThat ( e . getMessage ( ) ) |
deveObterValorICMSInterestadualRemetenteComoFoiSetado ( ) { final com . fincatto . documentofiscal . nfe310 . classes . nota . NFNotaInfoItemImpostoICMSUFDestino icmsUFDestino = new com . fincatto . documentofiscal . nfe310 . classes . nota . NFNotaInfoItemImpostoICMSUFDestino ( ) ; final java . lang . String valorICMSInterestadualRemetente = "9999999999999.99" ; icmsUFDestino . setValorICMSInterestadualRemetente ( new java . math . BigDecimal ( valorICMSInterestadualRemetente ) ) ; "<AssertPlaceHolder>" ; } getValorICMSInterestadualRemetente ( ) { return this . valorICMSInterestadualRemetente ; } | org . junit . Assert . assertEquals ( valorICMSInterestadualRemetente , icmsUFDestino . getValorICMSInterestadualRemetente ( ) ) |
testInvalidParse ( ) { final java . lang . String urn = "urn:ogcx:def:CRS:EPSG:6.8:4326" ; try { org . geotools . referencing . factory . URN_Parser . buildParser ( urn ) ; org . junit . Assert . fail ( ) ; } catch ( org . opengis . referencing . NoSuchAuthorityCodeException e ) { "<AssertPlaceHolder>" ; } } getAuthorityCode ( ) { return code ; } | org . junit . Assert . assertEquals ( urn , e . getAuthorityCode ( ) ) |
ConcurrentJobsStatusTooManyRequestsException ( ) { try { org . apache . hive . hcatalog . templeton . JobRunnable jobRunnable = ConcurrentJobsStatus ( 6 , org . apache . hive . hcatalog . templeton . TestConcurrentJobRequestsThreadsAndTimeout . config , false , false , statusJobHelper . getDelayedResonseAnswer ( 4 , org . apache . hive . hcatalog . templeton . TestConcurrentJobRequestsThreadsAndTimeout . statusBean ) ) ; verifyTooManyRequestsException ( jobRunnable . exception , this . statusTooManyRequestsExceptionMessage ) ; } catch ( java . lang . Exception e ) { "<AssertPlaceHolder>" ; } } verifyTooManyRequestsException ( java . lang . Throwable , java . lang . String ) { org . junit . Assert . assertTrue ( ( exception != null ) ) ; org . junit . Assert . assertTrue ( ( exception instanceof org . apache . hive . hcatalog . templeton . TooManyRequestsException ) ) ; org . apache . hive . hcatalog . templeton . TooManyRequestsException ex = ( ( org . apache . hive . hcatalog . templeton . TooManyRequestsException ) ( exception ) ) ; org . junit . Assert . assertTrue ( ( ( ex . httpCode ) == ( TooManyRequestsException . TOO_MANY_REQUESTS_429 ) ) ) ; org . junit . Assert . assertTrue ( exception . getMessage ( ) . contains ( expectedMessage ) ) ; } | org . junit . Assert . assertTrue ( false ) |
testStoreOverwriteExisting_shouldOverwrite ( ) { java . lang . String differentValue = "secondValue" ; org . apache . directory . shared . ldap . model . entry . Entry testEntry = newTestEntry ( testEntryName , baseDn ) ; org . apache . directory . shared . ldap . model . entry . Entry updatedEntry = newTestEntry ( testEntry . getDn ( ) ) ; updatedEntry . add ( ouAttribute , differentValue ) ; connection . add ( testEntry ) ; dao . storeOverwriteExisting ( updatedEntry ) ; org . apache . directory . shared . ldap . model . entry . Entry newResult = connection . lookup ( testEntry . getDn ( ) ) ; "<AssertPlaceHolder>" ; } contains ( java . util . List , org . apache . directory . shared . ldap . model . entry . Entry ) { for ( org . openengsb . infrastructure . ldap . model . Node n : nodes ) { if ( n . getEntry ( ) . equals ( entry ) ) { return true ; } } return false ; } | org . junit . Assert . assertThat ( newResult . contains ( ouAttribute , differentValue ) , org . hamcrest . Matchers . is ( true ) ) |
testModifyTicketWriteFileNullTicketId ( ) { org . irods . jargon . ticket . packinstr . TicketAdminInp pi = org . irods . jargon . ticket . packinstr . TicketAdminInp . instanceForModifyFileWriteNumber ( null , new java . lang . Integer ( 20000 ) ) ; "<AssertPlaceHolder>" ; } instanceForModifyFileWriteNumber ( java . lang . String , java . lang . Integer ) { if ( ( ticketId == null ) || ( ticketId . isEmpty ( ) ) ) { throw new java . lang . IllegalArgumentException ( "null<sp>or<sp>empty<sp>ticket<sp>id" ) ; } if ( numberOfFileWrites < 0 ) { throw new java . lang . IllegalArgumentException ( "illegal<sp>integer<sp>for<sp>write-file<sp>-<sp>must<sp>be<sp>0<sp>or<sp>greater" ) ; } return new org . irods . jargon . ticket . packinstr . TicketAdminInp ( TICKET_ADMIN_INP_API_NBR , "mod" , ticketId , "write-file" , numberOfFileWrites . toString ( ) , BLANK , BLANK ) ; } | org . junit . Assert . assertNotNull ( pi ) |
testCheckAndMaintainErrorInMaintenanceRunningVms ( ) { when ( vmInstanceDao . listByHostId ( com . cloud . resource . ResourceManagerImplTest . hostId ) ) . thenReturn ( java . util . Arrays . asList ( vm1 , vm2 ) ) ; boolean enterMaintenanceMode = resourceManager . checkAndMaintain ( com . cloud . resource . ResourceManagerImplTest . hostId ) ; verify ( resourceManager ) . isHostInMaintenance ( host , java . util . Arrays . asList ( vm1 , vm2 ) , new java . util . ArrayList ( ) , new java . util . ArrayList ( ) ) ; "<AssertPlaceHolder>" ; } verify ( org . apache . cloudstack . network . contrail . model . ModelController ) { assert _initialized : "initialized<sp>is<sp>false" ; assert ( _uuid ) != null : "uuid<sp>is<sp>not<sp>set" ; net . juniper . contrail . api . ApiConnector api = controller . getApiAccessor ( ) ; try { _vm = ( ( net . juniper . contrail . api . types . VirtualMachine ) ( api . findById ( net . juniper . contrail . api . types . VirtualMachine . class , _uuid ) ) ) ; } catch ( java . io . IOException e ) { org . apache . cloudstack . network . contrail . model . VirtualMachineModel . s_logger . error ( "virtual-machine<sp>verify" , e ) ; } if ( ( _vm ) == null ) { return false ; } for ( org . apache . cloudstack . network . contrail . model . ModelObject successor : successors ( ) ) { if ( ! ( successor . verify ( controller ) ) ) { return false ; } } return true ; } | org . junit . Assert . assertFalse ( enterMaintenanceMode ) |
testGetSenseDocument ( ) { org . wikidata . wdtk . datamodel . helpers . SenseDocument o1 = org . wikidata . wdtk . datamodel . helpers . Datamodel . makeSenseDocument ( factory . getSenseIdValue ( "L42-S1" , "foo" ) , java . util . Collections . singletonList ( factory . getMonolingualTextValue ( "en" , "foo" ) ) , java . util . Collections . emptyList ( ) ) ; org . wikidata . wdtk . datamodel . helpers . SenseDocument o2 = factory . getSenseDocument ( factory . getSenseIdValue ( "L42-S1" , "foo" ) , java . util . Collections . singletonList ( factory . getMonolingualTextValue ( "en" , "foo" ) ) , java . util . Collections . emptyList ( ) , 0 ) ; "<AssertPlaceHolder>" ; } getMonolingualTextValue ( java . lang . String , java . lang . String ) { return new org . wikidata . wdtk . datamodel . implementation . MonolingualTextValueImpl ( text , languageCode ) ; } | org . junit . Assert . assertEquals ( o1 , o2 ) |
createDiagnosticsStatusNotificationRequest_withStatus_returnsDiagnosticsStatusNotificationRequestWithStatus ( ) { eu . chargetime . ocpp . feature . profile . test . DiagnosticsStatus status = DiagnosticsStatus . Idle ; eu . chargetime . ocpp . feature . profile . test . DiagnosticsStatusNotificationRequest result = profile . createDiagnosticsStatusNotificationRequest ( status ) ; "<AssertPlaceHolder>" ; } getStatus ( ) { return status ; } | org . junit . Assert . assertThat ( result . getStatus ( ) , org . hamcrest . core . Is . is ( status ) ) |
testIsDelayedQueue_Success ( ) { final redis . clients . jedis . Jedis jedis = net . greghaines . jesque . TestUtils . createJedis ( net . greghaines . jesque . utils . TestJedisUtils . CONFIG ) ; jedis . zadd ( net . greghaines . jesque . utils . TestJedisUtils . TEST_KEY , 1.0 , "bar" ) ; "<AssertPlaceHolder>" ; } isDelayedQueue ( redis . clients . jedis . Jedis , java . lang . String ) { return net . greghaines . jesque . utils . JedisUtils . ZSET . equalsIgnoreCase ( jedis . type ( key ) ) ; } | org . junit . Assert . assertTrue ( net . greghaines . jesque . utils . JedisUtils . isDelayedQueue ( jedis , net . greghaines . jesque . utils . TestJedisUtils . TEST_KEY ) ) |
testFetchPlan1 ( ) { com . orientechnologies . orient . core . sql . parser . SimpleNode result = checkRightSyntax ( ( "" + "select<sp>'Ay'<sp>as<sp>a<sp>,<sp>'bEE'<sp>as<sp>b<sp>from<sp>Foo<sp>fetchplan<sp>*:1" ) ) ; "<AssertPlaceHolder>" ; com . orientechnologies . orient . core . sql . parser . OSelectStatement select = ( ( com . orientechnologies . orient . core . sql . parser . OSelectStatement ) ( result ) ) ; } checkRightSyntax ( java . lang . String ) { com . orientechnologies . orient . core . sql . parser . SimpleNode result = checkSyntax ( query , true ) ; return checkSyntax ( result . toString ( ) , true ) ; } | org . junit . Assert . assertTrue ( ( result instanceof com . orientechnologies . orient . core . sql . parser . OSelectStatement ) ) |
decode ( ) { final java . io . ByteArrayInputStream stream = new java . io . ByteArrayInputStream ( encoded ) ; final com . flagstone . transform . coder . SWFDecoder decoder = new com . flagstone . transform . coder . SWFDecoder ( stream ) ; fixture = new com . flagstone . transform . sound . Envelope ( decoder ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertNotNull ( fixture ) |
copyConstructor_worksOnSafeCopy ( ) { org . eclipse . rap . json . JsonObject copy = new org . eclipse . rap . json . JsonObject ( object ) ; object . add ( "foo" , 23 ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( ( childCollections . size ( ) ) == 0 ) && ( ( cheatsheets . size ( ) ) == 0 ) ; } | org . junit . Assert . assertTrue ( copy . isEmpty ( ) ) |
testNotNullDefault ( ) { com . j256 . ormlite . dao . Dao < com . j256 . ormlite . dao . JdbcBaseDaoImplTest . NotNullDefault , java . lang . Object > dao = createDao ( com . j256 . ormlite . dao . JdbcBaseDaoImplTest . NotNullDefault . class , true ) ; com . j256 . ormlite . dao . JdbcBaseDaoImplTest . NotNullDefault notNullDefault = new com . j256 . ormlite . dao . JdbcBaseDaoImplTest . NotNullDefault ( ) ; "<AssertPlaceHolder>" ; } createDao ( java . lang . Class , boolean ) { if ( ( connectionSource ) == null ) { throw new java . sql . SQLException ( com . j256 . ormlite . BaseJdbcTest . DATASOURCE_ERROR ) ; } @ com . j256 . ormlite . SuppressWarnings ( "unchecked" ) com . j256 . ormlite . dao . BaseDaoImpl < T , ID > dao = ( ( com . j256 . ormlite . dao . BaseDaoImpl < T , ID > ) ( com . j256 . ormlite . dao . DaoManager . createDao ( connectionSource , clazz ) ) ) ; return configDao ( dao , createTable ) ; } | org . junit . Assert . assertEquals ( 1 , dao . create ( notNullDefault ) ) |
shouldDetectCommonsAndUberJar ( ) { org . infinispan . commons . util . uberjar . ManifestUberJarDuplicatedJarsWarner scanner = new org . infinispan . commons . util . uberjar . ManifestUberJarDuplicatedJarsWarner ( ) { @ org . infinispan . commons . util . uberjar . Override java . util . List < java . lang . String > getBundleSymbolicNames ( ) { return java . util . Arrays . asList ( "org.infinispan.embedded" , "org.infinispan.commons" ) ; } } ; java . lang . Boolean isClasspathCorrect = scanner . isClasspathCorrect ( ) ; "<AssertPlaceHolder>" ; } isClasspathCorrect ( ) { java . util . List < java . lang . String > bundleNames = getBundleSymbolicNames ( ) ; long numberOfMatches = bundleNames . stream ( ) . filter ( hasRemoteUberJar ( ) . or ( hasEmbeddedUberJar ( ) ) . or ( hasCommons ( ) ) ) . count ( ) ; return numberOfMatches < 2 ; } | org . junit . Assert . assertFalse ( isClasspathCorrect ) |
testSerialization ( ) { org . jfree . chart . renderer . category . CategoryStepRenderer r1 = new org . jfree . chart . renderer . category . CategoryStepRenderer ( ) ; org . jfree . chart . renderer . category . CategoryStepRenderer r2 = ( ( org . jfree . chart . renderer . category . CategoryStepRenderer ) ( org . jfree . chart . TestUtilities . serialised ( r1 ) ) ) ; "<AssertPlaceHolder>" ; } serialised ( java . lang . Object ) { java . lang . Object result = null ; java . io . ByteArrayOutputStream buffer = new java . io . ByteArrayOutputStream ( ) ; java . io . ObjectOutput out ; try { out = new java . io . ObjectOutputStream ( buffer ) ; out . writeObject ( original ) ; out . close ( ) ; java . io . ObjectInput in = new java . io . ObjectInputStream ( new java . io . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; result = in . readObject ( ) ; in . close ( ) ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } catch ( java . lang . ClassNotFoundException e ) { throw new java . lang . RuntimeException ( e ) ; } return result ; } | org . junit . Assert . assertEquals ( r1 , r2 ) |
createNewDrug_shouldCreateNewDrug ( ) { int before = service . getAllDrugs ( true ) . size ( ) ; java . lang . String json = "{\"concept\":\"0cbe2ed3-cd5f-4f46-9459-26127c9265ab\",\"name\":\"New<sp>Drug<sp>name\",\"dosageForm\":\"0cbe2ed3-cd5f-4f46-9459-26127c9265ab\",\"minimumDailyDose\":\"10\",\"maximumDailyDose\":\"100\",\"units\":\"mg\"}" ; org . openmrs . module . webservices . rest . SimpleObject post = new org . codehaus . jackson . map . ObjectMapper ( ) . readValue ( json , org . openmrs . module . webservices . rest . SimpleObject . class ) ; java . lang . Object drug = controller . createNewDrug ( post , request , response ) ; "<AssertPlaceHolder>" ; } createNewDrug ( org . openmrs . module . webservices . rest . SimpleObject , javax . servlet . http . HttpServletRequest , javax . servlet . http . HttpServletResponse ) { initDrugController ( ) ; java . lang . String conceptUuid = post . get ( "concept" ) . toString ( ) ; org . openmrs . Concept concept = service . getConceptByUuid ( conceptUuid ) ; if ( concept == null ) { throw new org . openmrs . module . webservices . rest . web . response . ObjectNotFoundException ( ) ; } org . openmrs . Drug drug = new org . openmrs . Drug ( ) ; drug . setConcept ( concept ) ; updateDrugFieldsFromPostData ( drug , post ) ; org . openmrs . Drug drugJustCreated = service . saveDrug ( drug ) ; if ( ( post . get ( "drugInfo" ) ) != null ) { this . createNewDrugInfo ( drugJustCreated , ( ( java . util . LinkedHashMap ) ( post . get ( "drugInfo" ) ) ) ) ; } return org . openmrs . module . webservices . rest . web . RestUtil . created ( response , getDrugAsSimpleObject ( drugJustCreated ) ) ; } | org . junit . Assert . assertEquals ( ( before + 1 ) , service . getAllDrugs ( true ) . size ( ) ) |
testVerticalList ( ) { org . antlr . test . TreeAdaptor adaptor = new org . antlr . test . CommonTreeAdaptor ( ) ; org . antlr . test . TreeWizard wiz = new org . antlr . test . TreeWizard ( adaptor , org . antlr . test . TestTreeIterator . tokens ) ; org . antlr . test . CommonTree t = ( ( org . antlr . test . CommonTree ) ( wiz . create ( "(A<sp>(B<sp>C))" ) ) ) ; org . antlr . test . TreeIterator it = new org . antlr . test . TreeIterator ( t ) ; java . lang . StringBuffer buf = org . antlr . test . TestTreeIterator . toString ( it ) ; java . lang . String expecting = "A<sp>DOWN<sp>B<sp>DOWN<sp>C<sp>UP<sp>UP<sp>EOF" ; java . lang . String found = buf . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( ( ( "[" + ( grammar . name ) ) + "." ) + ( name ) ) + ",index=" ) + ( index ) ) + ",line=" ) + ( tree . getToken ( ) . getLine ( ) ) ) + "]" ; } | org . junit . Assert . assertEquals ( expecting , found ) |
testThatDoesntNormalizeIfStartsWithHttp ( ) { java . lang . String url = new net . vidageek . crawler . component . DefaultLinkNormalizer ( "http://test.com/foo" ) . normalize ( "http://other.com/bar" ) ; "<AssertPlaceHolder>" ; } normalize ( java . lang . String ) { java . lang . String normalizedUrl = delegate . normalize ( url ) . toLowerCase ( ) ; if ( normalizedUrl . contains ( "#" ) ) return normalizedUrl . substring ( 0 , normalizedUrl . indexOf ( "#" ) ) ; return normalizedUrl ; } | org . junit . Assert . assertEquals ( "http://other.com/bar" , url ) |
testUpdate ( ) { try ( com . syncleus . ferma . tx . Tx tx = tx ( ) ) { com . gentics . mesh . core . data . TagFamily tagFamily = tagFamily ( "colors" ) ; tagFamily . setName ( "new<sp>Name" ) ; "<AssertPlaceHolder>" ; } } getName ( ) { return "Fix<sp>data<sp>inconsistency<sp>for<sp>older<sp>versions." ; } | org . junit . Assert . assertEquals ( "new<sp>Name" , tagFamily . getName ( ) ) |
serialize ( ) { com . google . gson . Gson gson = com . github . seratch . jslack . common . json . GsonFactory . createSnakeCase ( ) ; com . github . seratch . jslack . api . model . event . ChannelCreatedEvent event = new com . github . seratch . jslack . api . model . event . ChannelCreatedEvent ( ) ; event . setChannel ( new com . github . seratch . jslack . api . model . event . ChannelCreatedEvent . Channel ( ) ) ; java . lang . String generatedJson = gson . toJson ( event ) ; java . lang . String expectedJson = "{\"type\":\"channel_created\",\"channel\":{}}" ; "<AssertPlaceHolder>" ; } createSnakeCase ( ) { return new com . google . gson . GsonBuilder ( ) . setFieldNamingPolicy ( FieldNamingPolicy . LOWER_CASE_WITH_UNDERSCORES ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . LayoutBlock . class , new com . github . seratch . jslack . common . json . GsonLayoutBlockFactory ( ) ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . composition . TextObject . class , new com . github . seratch . jslack . common . json . GsonTextObjectFactory ( ) ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . ContextBlockElement . class , new com . github . seratch . jslack . common . json . GsonContextBlockElementFactory ( ) ) . registerTypeAdapter ( com . github . seratch . jslack . api . model . block . element . BlockElement . class , new com . github . seratch . jslack . common . json . GsonBlockElementFactory ( ) ) . create ( ) ; } | org . junit . Assert . assertThat ( generatedJson , org . hamcrest . CoreMatchers . is ( expectedJson ) ) |
testBoolean2 ( ) { java . lang . Boolean convertedBooolean = _converter . convertType ( ExpandoColumnConstants . BOOLEAN , "false" ) ; "<AssertPlaceHolder>" ; } convertType ( int , java . lang . Object ) { if ( data == null ) { return ( ( T ) ( data ) ) ; } data = handleCollections ( type , data ) ; data = handleStrings ( type , data ) ; if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . BOOLEAN ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , Boolean . TYPE ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . BOOLEAN_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , boolean [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . DATE ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , java . util . Date . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . DATE_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , java . util . Date [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . DOUBLE ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , Double . TYPE ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . DOUBLE_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , double [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . FLOAT ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , Float . TYPE ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . FLOAT_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , float [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . INTEGER ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , Integer . TYPE ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . INTEGER_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , int [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . LONG ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , Long . TYPE ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . LONG_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , long [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . NUMBER ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , com . liferay . portlet . expando . service . impl . Number . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . NUMBER_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , com . liferay . portlet . expando . service . impl . Number [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . SHORT ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , Short . TYPE ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . SHORT_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , short [ ] . class ) ; } else if ( type == ( com . liferay . expando . kernel . model . ExpandoColumnConstants . STRING_ARRAY ) ) { data = jodd . typeconverter . TypeConverterManager . convertType ( data , java . lang . String [ ] . class ) ; } return ( ( T ) ( data ) ) ; } | org . junit . Assert . assertFalse ( convertedBooolean ) |
testRoutedSimpleExtraction ( ) { final org . opendaylight . controller . md . sal . binding . impl . ContextReferenceExtractor extractor = org . opendaylight . controller . md . sal . binding . impl . ContextReferenceExtractor . from ( org . opendaylight . yang . gen . v1 . urn . opendaylight . params . xml . ns . yang . controller . md . sal . test . rpc . routing . rev140701 . RoutedSimpleRouteInput . class ) ; final org . opendaylight . yang . gen . v1 . urn . opendaylight . params . xml . ns . yang . controller . md . sal . test . rpc . routing . rev140701 . RoutedSimpleRouteInput input = new org . opendaylight . yang . gen . v1 . urn . opendaylight . params . xml . ns . yang . controller . md . sal . test . rpc . routing . rev140701 . RoutedSimpleRouteInputBuilder ( ) . setRoute ( org . opendaylight . controller . md . sal . binding . impl . ContextExtractorTest . TEST_ROUTE ) . build ( ) ; final org . opendaylight . yangtools . yang . binding . InstanceIdentifier < ? > extractedValue = extractor . extract ( input ) ; "<AssertPlaceHolder>" ; } extract ( org . opendaylight . yangtools . yang . binding . DataObject ) { try { return ( ( org . opendaylight . yangtools . yang . binding . InstanceIdentifier < ? > ) ( handle . invokeExact ( obj ) ) ) ; } catch ( java . lang . Throwable e ) { throw com . google . common . base . Throwables . propagate ( e ) ; } } | org . junit . Assert . assertSame ( org . opendaylight . controller . md . sal . binding . impl . ContextExtractorTest . TEST_ROUTE , extractedValue ) |
testGetWorkerAddress ( ) { "<AssertPlaceHolder>" ; } getWorkerAddress ( ) { if ( ( workerAddress ) == null ) { return null ; } return com . hazelcast . simulator . protocol . core . SimulatorAddress . fromString ( workerAddress ) ; } | org . junit . Assert . assertEquals ( workerAddress , operation . getWorkerAddress ( ) ) |
testToFullyQualifiedNameClassString ( ) { final java . lang . String expected = "org.apache.commons.lang3.Test.properties" ; final java . lang . String actual = org . apache . commons . lang3 . ClassPathUtils . toFullyQualifiedName ( org . apache . commons . lang3 . ClassPathUtils . class , "Test.properties" ) ; "<AssertPlaceHolder>" ; } toFullyQualifiedName ( java . lang . Class , java . lang . String ) { org . apache . commons . lang3 . Validate . notNull ( context , "Parameter<sp>'%s'<sp>must<sp>not<sp>be<sp>null!" , "context" ) ; org . apache . commons . lang3 . Validate . notNull ( resourceName , "Parameter<sp>'%s'<sp>must<sp>not<sp>be<sp>null!" , "resourceName" ) ; return org . apache . commons . lang3 . ClassPathUtils . toFullyQualifiedName ( context . getPackage ( ) , resourceName ) ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testBasicVisibility ( ) { org . apache . accumulo . core . client . summary . SummarizerConfiguration sc = org . apache . accumulo . core . client . summary . SummarizerConfiguration . builder ( org . apache . accumulo . core . client . summary . summarizers . EntryLengthSummarizer . class ) . build ( ) ; org . apache . accumulo . core . client . summary . summarizers . EntryLengthSummarizer entrySum = new org . apache . accumulo . core . client . summary . summarizers . EntryLengthSummarizer ( ) ; org . apache . accumulo . core . data . Key k1 = new org . apache . accumulo . core . data . Key ( "r1" , "row.sum" 9 , "qualifier.logHist.1" 2 , "value.min" 2 ) ; org . apache . accumulo . core . data . Key k2 = new org . apache . accumulo . core . data . Key ( "r2" , "qualifier.logHist.1" 4 , "value.min" 6 , "row.sum" 4 ) ; org . apache . accumulo . core . data . Key k3 = new org . apache . accumulo . core . data . Key ( "row.sum" 1 , "row.sum" 8 , "q3" , "qualifier.logHist.1" 7 ) ; org . apache . accumulo . core . client . summary . Summarizer . Collector collector = entrySum . collector ( sc ) ; collector . accept ( k1 , new org . apache . accumulo . core . data . Value ( "row.sum" 3 ) ) ; collector . accept ( k2 , new org . apache . accumulo . core . data . Value ( "row.sum" 3 ) ) ; collector . accept ( k3 , new org . apache . accumulo . core . data . Value ( "row.sum" 3 ) ) ; java . util . HashMap < java . lang . String , java . lang . Long > stats = new java . util . HashMap ( ) ; collector . summarize ( stats :: put ) ; java . util . HashMap < java . lang . String , java . lang . Long > expected = new java . util . HashMap ( ) ; expected . put ( "key.min" , 8L ) ; expected . put ( "row.sum" 6 , 8L ) ; expected . put ( "qualifier.logHist.1" 9 , 24L ) ; expected . put ( "value.min" 5 , 3L ) ; expected . put ( "row.sum" 7 , 2L ) ; expected . put ( "qualifier.logHist.1" 8 , 2L ) ; expected . put ( "row.sum" , 6L ) ; expected . put ( "value.min" 4 , 3L ) ; expected . put ( "row.sum" 5 , 2L ) ; expected . put ( "value.min" 0 , 2L ) ; expected . put ( "family.sum" , 6L ) ; expected . put ( "qualifier.logHist.1" 6 , 3L ) ; expected . put ( "qualifier.min" , 2L ) ; expected . put ( "qualifier.logHist.1" 5 , 2L ) ; expected . put ( "value.min" 3 , 6L ) ; expected . put ( "qualifier.logHist.1" , 3L ) ; expected . put ( "value.min" 7 , 2L ) ; expected . put ( "qualifier.logHist.1" 0 , 2L ) ; expected . put ( "qualifier.logHist.1" 1 , 6L ) ; expected . put ( "qualifier.logHist.1" 3 , 3L ) ; expected . put ( "value.min" , 0L ) ; expected . put ( "value.min" 1 , 0L ) ; expected . put ( "row.sum" 2 , 0L ) ; expected . put ( "value.logHist.0" , 3L ) ; expected . put ( "row.sum" 0 , 3L ) ; "<AssertPlaceHolder>" ; } put ( java . lang . Long , org . apache . accumulo . server . tabletserver . LargestFirstMemoryManager$TabletInfo ) { if ( ( map . size ( ) ) == ( max ) ) { if ( ( key . compareTo ( map . firstKey ( ) ) ) < 0 ) return false ; try { add ( key , value ) ; return true ; } finally { map . remove ( map . firstKey ( ) ) ; } } else { add ( key , value ) ; return true ; } } | org . junit . Assert . assertEquals ( expected , stats ) |
testUpdateTagTypeNoChangesToDisplayNameExceptForCase ( ) { tagTypeDaoTestHelper . createTagTypeEntity ( org . finra . herd . service . TAG_TYPE , org . finra . herd . service . TAG_TYPE_DISPLAY_NAME , org . finra . herd . service . INTEGER_VALUE , org . finra . herd . service . TAG_TYPE_DESCRIPTION ) ; org . finra . herd . model . api . xml . TagType updatedTagType = tagTypeService . updateTagType ( new org . finra . herd . model . api . xml . TagTypeKey ( TAG_TYPE ) , new org . finra . herd . model . api . xml . TagTypeUpdateRequest ( org . finra . herd . service . TAG_TYPE_DISPLAY_NAME . toLowerCase ( ) , INTEGER_VALUE_2 , TAG_TYPE_DESCRIPTION_2 ) ) ; "<AssertPlaceHolder>" ; } createTagTypeEntity ( java . lang . String , java . lang . String , java . lang . Integer , java . lang . String ) { org . finra . herd . model . jpa . TagTypeEntity tagTypeEntity = new org . finra . herd . model . jpa . TagTypeEntity ( ) ; tagTypeEntity . setCode ( typeCode ) ; tagTypeEntity . setOrderNumber ( orderNumber ) ; tagTypeEntity . setDisplayName ( displayName ) ; tagTypeEntity . setDescription ( description ) ; return tagTypeDao . saveAndRefresh ( tagTypeEntity ) ; } | org . junit . Assert . assertEquals ( new org . finra . herd . model . api . xml . TagType ( new org . finra . herd . model . api . xml . TagTypeKey ( TAG_TYPE ) , org . finra . herd . service . TAG_TYPE_DISPLAY_NAME . toLowerCase ( ) , INTEGER_VALUE_2 , TAG_TYPE_DESCRIPTION_2 ) , updatedTagType ) |
testInitConfFromServletContext ( ) { com . amadeus . session . SessionConfiguration sc = new com . amadeus . session . SessionConfiguration ( ) ; when ( servletContext . getAttribute ( SessionHelpers . SESSION_CONFIGURATION ) ) . thenReturn ( sc ) ; com . amadeus . session . SessionConfiguration configuration = com . amadeus . session . servlet . SessionHelpers . initConf ( servletContext ) ; "<AssertPlaceHolder>" ; verify ( servletContext , never ( ) ) . setAttribute ( org . mockito . Mockito . eq ( SessionHelpers . SESSION_CONFIGURATION ) , any ( ) ) ; } initConf ( javax . servlet . ServletContext ) { com . amadeus . session . SessionConfiguration sessionConfiguration = ( ( com . amadeus . session . SessionConfiguration ) ( context . getAttribute ( com . amadeus . session . servlet . SessionHelpers . SESSION_CONFIGURATION ) ) ) ; if ( sessionConfiguration == null ) { sessionConfiguration = new com . amadeus . session . SessionConfiguration ( ) ; context . setAttribute ( com . amadeus . session . servlet . SessionHelpers . SESSION_CONFIGURATION , sessionConfiguration ) ; com . amadeus . session . servlet . WebXmlParser . parseWebXml ( sessionConfiguration , context ) ; sessionConfiguration . initializeFrom ( new com . amadeus . session . SessionConfiguration . AttributeProvider ( ) { @ com . amadeus . session . servlet . Override public java . lang . String getAttribute ( java . lang . String key ) { return context . getInitParameter ( key ) ; } @ com . amadeus . session . servlet . Override public java . lang . Object source ( ) { return context . getContextPath ( ) ; } } ) ; if ( ( sessionConfiguration . getTrueNamespace ( ) ) == null ) { sessionConfiguration . setNamespace ( context . getContextPath ( ) ) ; } } return sessionConfiguration ; } | org . junit . Assert . assertEquals ( sc , configuration ) |
shouldOrderCanBeClosedWhenTypeIsForEachOpAndRecordIsLastAndThereIsEnoughtLastRecords ( ) { orderHasEnabledAutoClose ( ) ; stubTypeOfProductionRecording ( TypeOfProductionRecording . FOR_EACH ) ; productionTrackingIsLast ( ) ; stubSearchCriteriaResults ( 1L , 2L , 3L ) ; boolean shouldClose = orderClosingHelper . orderShouldBeClosed ( productionTracking ) ; "<AssertPlaceHolder>" ; } orderShouldBeClosed ( com . qcadoo . model . api . Entity ) { com . qcadoo . model . api . Entity order = productionTracking . getBelongsToField ( ProductionTrackingFields . ORDER ) ; if ( order == null ) { return false ; } java . lang . Boolean autoCloseOrder = order . getBooleanField ( OrderFieldsPC . AUTO_CLOSE_ORDER ) ; java . lang . Boolean isLastRecord = productionTracking . getBooleanField ( ProductionTrackingFields . LAST_TRACKING ) ; com . qcadoo . mes . productionCounting . constants . TypeOfProductionRecording recType = com . qcadoo . mes . productionCounting . constants . TypeOfProductionRecording . parseString ( order . getStringField ( OrderFieldsPC . TYPE_OF_PRODUCTION_RECORDING ) ) ; return ( isLastRecord && autoCloseOrder ) && ( ( ! ( TypeOfProductionRecording . FOR_EACH . equals ( recType ) ) ) || ( eachOperationHasLastRecords ( order , productionTracking ) ) ) ; } | org . junit . Assert . assertTrue ( shouldClose ) |
load_changeRefreshTimeInLoader_triggeredViaEntryProcessor ( ) { final long _probeTime = 4711 ; org . cache2k . Cache < java . lang . Integer , java . lang . Integer > c = target . cache ( new CacheRule . Specialization < java . lang . Integer , java . lang . Integer > ( ) { @ java . lang . Override public void extend ( final Cache2kBuilder < java . lang . Integer , java . lang . Integer > b ) { b . recordRefreshedTime ( true ) . wrappingLoader ( new AdvancedCacheLoader < java . lang . Integer , LoadDetail < java . lang . Integer > > ( ) { @ java . lang . Override public LoadDetail < java . lang . Integer > load ( final java . lang . Integer key , final long currentTime , final CacheEntry < java . lang . Integer , LoadDetail < java . lang . Integer > > currentEntry ) throws java . lang . Exception { return org . cache2k . integration . Loaders . wrapRefreshedTime ( key , _probeTime ) ; } } ) ; } } ) ; c . invoke ( 1 , new org . cache2k . processor . EntryProcessor < java . lang . Integer , java . lang . Integer , java . lang . Object > ( ) { @ org . cache2k . test . core . Override public java . lang . Object process ( final org . cache2k . processor . MutableCacheEntry < java . lang . Integer , java . lang . Integer > e ) { java . lang . Integer v = e . getValue ( ) ; "<AssertPlaceHolder>" ; return null ; } } ) ; } getRefreshedTime ( ) { if ( ( refreshTime ) != ( NEUTRAL ) ) { return refreshTime ; } return originalExists ? entry . getRefreshTime ( ) : 0 ; } | org . junit . Assert . assertEquals ( _probeTime , e . getRefreshedTime ( ) ) |
getting_user_should_not_return_user_if_there_is_no_preexisting_http_session ( ) { javax . servlet . http . HttpServletRequest request = mock ( javax . servlet . http . HttpServletRequest . class ) ; when ( request . getSession ( false ) ) . thenReturn ( null ) ; com . commafeed . frontend . session . SessionHelper sessionHelper = new com . commafeed . frontend . session . SessionHelper ( request ) ; java . util . Optional < com . commafeed . backend . model . User > user = sessionHelper . getLoggedInUser ( ) ; "<AssertPlaceHolder>" ; } getLoggedInUser ( ) { java . util . Optional < javax . servlet . http . HttpSession > session = getSession ( false ) ; if ( session . isPresent ( ) ) { com . commafeed . backend . model . User user = ( ( com . commafeed . backend . model . User ) ( session . get ( ) . getAttribute ( com . commafeed . frontend . session . SessionHelper . SESSION_KEY_USER ) ) ) ; return java . util . Optional . ofNullable ( user ) ; } return java . util . Optional . empty ( ) ; } | org . junit . Assert . assertFalse ( user . isPresent ( ) ) |
testExpressionFactoryMatchAny ( ) { createDataSet ( ) ; org . apache . cayenne . query . SelectQuery < org . apache . cayenne . testdo . mt . ClientMtTable2 > table2Query = new org . apache . cayenne . query . SelectQuery ( org . apache . cayenne . testdo . mt . ClientMtTable2 . class ) ; table2Query . addOrdering ( new org . apache . cayenne . query . Ordering ( "db:TABLE2_ID" , SortOrder . ASCENDING ) ) ; java . util . List < org . apache . cayenne . testdo . mt . ClientMtTable2 > table2List = context . select ( table2Query ) ; org . apache . cayenne . testdo . mt . ClientMtTable2 element_1 = table2List . get ( 0 ) ; org . apache . cayenne . testdo . mt . ClientMtTable2 element_2 = table2List . get ( 10 ) ; org . apache . cayenne . exp . Expression exp = org . apache . cayenne . exp . ExpressionFactory . matchAnyExp ( element_1 , element_2 ) ; table2Query = new org . apache . cayenne . query . SelectQuery ( org . apache . cayenne . testdo . mt . ClientMtTable2 . class , exp ) ; table2List = context . select ( table2Query ) ; "<AssertPlaceHolder>" ; } size ( ) { return neighbors . size ( ) ; } | org . junit . Assert . assertEquals ( 2 , table2List . size ( ) ) |
SaveFile ( ) { cruise . umple . util . SampleFileWriter . createFile ( "myapp.ump" , "namespace<sp>Airline\nclass<sp>Airline{1<sp>--<sp>*<sp>RegularFlight;}class<sp>RegularFlight<sp>{}" ) ; cruise . umple . util . SampleFileWriter . createFile ( "myapps.txt" , "p1\nmyapp.ump\n" ) ; cruise . umple . stats . StatsMain . main ( new java . lang . String [ ] { "results.xls" , "myapps.txt" } ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( true , new cruise . umple . stats . File ( "results.xls" ) . exists ( ) ) |
testGetOutOfBandSchema ( ) { org . apache . avro . Schema mainSchema = org . apache . avro . SchemaBuilder . builder ( ) . record ( "Main" ) . fields ( ) . name ( "id" ) . type ( ) . intType ( ) . noDefault ( ) . endRecord ( ) ; org . apache . avro . Schema outOfBandSchema = org . apache . avro . SchemaBuilder . builder ( ) . record ( "EmptySchema" ) . fields ( ) . endRecord ( ) ; org . apache . avro . Schema rootSchema = org . apache . avro . SchemaBuilder . record ( "Root" ) . fields ( ) . name ( "Main" ) . type ( mainSchema ) . noDefault ( ) . name ( "OutOfBand" ) . type ( outOfBandSchema ) . noDefault ( ) . endRecord ( ) ; org . apache . avro . Schema actualOutOfBandSchema = org . talend . components . common . avro . RootSchemaUtils . getOutOfBandSchema ( rootSchema ) ; "<AssertPlaceHolder>" ; } getOutOfBandSchema ( org . apache . avro . Schema ) { if ( ! ( org . talend . components . common . avro . RootSchemaUtils . isRootSchema ( rootSchema ) ) ) { throw new java . lang . IllegalArgumentException ( "Input<sp>schema<sp>should<sp>be<sp>Root<sp>schema" ) ; } org . apache . avro . Schema outOfBandSchema = rootSchema . getField ( org . talend . components . common . avro . RootSchemaUtils . OUTOFBAND_FIELD_NAME ) . schema ( ) ; return outOfBandSchema ; } | org . junit . Assert . assertEquals ( outOfBandSchema , actualOutOfBandSchema ) |
serializeDeserialize ( ) { final byte [ ] keyBlobBytes = createKeyBlobBytes ( ) ; final com . amazonaws . encryptionsdk . model . KeyBlob reconstructedKeyBlob = deserialize ( keyBlobBytes ) ; final byte [ ] reconstructedKeyBlobBytes = reconstructedKeyBlob . toByteArray ( ) ; "<AssertPlaceHolder>" ; } toByteArray ( ) { if ( ( ( headerNonce_ ) == null ) || ( ( headerTag_ ) == null ) ) { throw new com . amazonaws . encryptionsdk . exception . AwsCryptoException ( "Header<sp>nonce<sp>and<sp>tag<sp>cannot<sp>be<sp>null." ) ; } final byte [ ] serializedFields = serializeAuthenticatedFields ( ) ; final int outLen = ( ( serializedFields . length ) + ( headerNonce_ . length ) ) + ( headerTag_ . length ) ; final java . nio . ByteBuffer serializedBytes = java . nio . ByteBuffer . allocate ( outLen ) ; serializedBytes . put ( serializedFields ) ; serializedBytes . put ( headerNonce_ ) ; serializedBytes . put ( headerTag_ ) ; return serializedBytes . array ( ) ; } | org . junit . Assert . assertArrayEquals ( reconstructedKeyBlobBytes , keyBlobBytes ) |
testContainsMaterialAmount ( ) { org . spout . api . inventory . ItemStack item = getRandomItem ( ) ; org . spout . api . material . Material mat = item . getMaterial ( ) ; int amount = item . getAmount ( ) ; inventory . clear ( ) ; inventory . add ( item ) ; "<AssertPlaceHolder>" ; } contains ( org . spout . api . material . Material , int ) { return ( getAmount ( material ) ) >= amount ; } | org . junit . Assert . assertTrue ( inventory . contains ( mat , amount ) ) |
testGetMetricsResources ( ) { java . util . Map < java . lang . String , java . lang . Object > report = reportLoader . loadReport ( "test" , "test.yaml" ) ; java . util . List < jp . primecloud . auto . puppet . report . MetricsResource > metricsResources = reportAnalyzer . getMetricsResources ( report ) ; for ( jp . primecloud . auto . puppet . report . MetricsResource metricsResource : metricsResources ) { log . trace ( org . apache . commons . lang . builder . ReflectionToStringBuilder . toString ( metricsResource , ToStringStyle . SHORT_PREFIX_STYLE ) ) ; } "<AssertPlaceHolder>" ; } getMetricsResources ( java . util . Map ) { java . util . List < jp . primecloud . auto . puppet . report . MetricsResource > metricsResources = new java . util . ArrayList < jp . primecloud . auto . puppet . report . MetricsResource > ( ) ; if ( report == null ) { return metricsResources ; } java . util . Map < java . lang . String , java . lang . Object > metrics = ( ( java . util . Map < java . lang . String , java . lang . Object > ) ( report . get ( "metrics" ) ) ) ; if ( metrics == null ) { return metricsResources ; } java . util . Map < java . lang . String , java . lang . Object > resources = ( ( java . util . Map < java . lang . String , java . lang . Object > ) ( metrics . get ( "resources" ) ) ) ; if ( resources == null ) { return metricsResources ; } java . util . List < java . util . List < java . lang . Object > > values = ( ( java . util . List < java . util . List < java . lang . Object > > ) ( resources . get ( "values" ) ) ) ; if ( values == null ) { return metricsResources ; } for ( java . util . List < java . lang . Object > value : values ) { java . lang . String name = value . get ( 1 ) . toString ( ) ; int count = java . lang . Integer . parseInt ( value . get ( 2 ) . toString ( ) ) ; jp . primecloud . auto . puppet . report . MetricsResource metricsResource = new jp . primecloud . auto . puppet . report . MetricsResource ( ) ; metricsResource . setName ( name ) ; metricsResource . setCount ( count ) ; metricsResources . add ( metricsResource ) ; } return metricsResources ; } | org . junit . Assert . assertEquals ( 8 , metricsResources . size ( ) ) |
testAddCardinalityFieldsToProjectFieldsNoWhitelist ( ) { datawave . query . cardinality . CardinalityConfiguration config = new datawave . query . cardinality . CardinalityConfiguration ( ) ; config . setCardinalityUidField ( "UUID" ) ; config . setCardinalityFieldReverseMapping ( reverseMap ) ; config . setCardinalityFields ( asSet ( new java . lang . String [ ] { "R_LABEL" , "QUERY_USER|PROTOCOL" } ) ) ; java . util . Set < java . lang . String > originalProjectFieldsSet = java . util . Collections . emptySet ( ) ; java . util . Set < java . lang . String > revisedProjectFields = config . getRevisedProjectFields ( QUERY_MODEL , originalProjectFieldsSet ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . typeMetadata . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , revisedProjectFields . size ( ) ) |
jsonOptional ( ) { com . treasuredata . client . model . TDUser user = com . treasuredata . client . model . ObjectMappers . compactMapper ( ) . readValue ( com . treasuredata . client . model . TDUserTest . USER_WITHOUT_OPTIONAL_FIELDS_JSON , com . treasuredata . client . model . TDUser . class ) ; java . lang . String serialized = com . treasuredata . client . model . ObjectMappers . compactMapper ( ) . writeValueAsString ( user ) ; com . treasuredata . client . model . TDUser parsed = com . treasuredata . client . model . ObjectMappers . compactMapper ( ) . readValue ( serialized , com . treasuredata . client . model . TDUser . class ) ; "<AssertPlaceHolder>" ; } compactMapper ( ) { return com . treasuredata . client . model . ObjectMappers . Lazy . COMPACT_MAPPER ; } | org . junit . Assert . assertThat ( parsed , org . hamcrest . Matchers . is ( user ) ) |
checkIfConnectionIsClosedWithoutTheConnection ( ) { "<AssertPlaceHolder>" ; verify ( connection , times ( 1 ) ) . isClosed ( ) ; } getConnectionToTest ( ) { return new com . arjuna . ats . internal . jdbc . ConnectionImple ( null , null , null , null , xaDataSource ) ; } | org . junit . Assert . assertFalse ( getConnectionToTest ( ) . isClosed ( ) ) |
testLoadImportConfigurationWithEmptyFile_shouldReturnEmptyList ( ) { org . apache . commons . io . FileUtils . forceDelete ( storageFile ) ; java . util . List < org . openengsb . core . api . model . ConfigItem < org . openengsb . core . workflow . drools . model . ImportDeclaration > > loaded = service . load ( new java . util . HashMap < java . lang . String , java . lang . String > ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return columns . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , loaded . size ( ) ) |
testInvalidTeacherStaffAssociation ( ) { org . slc . sli . api . security . context . validator . Set < java . lang . String > ids = new org . slc . sli . api . security . context . validator . HashSet < java . lang . String > ( org . slc . sli . api . security . context . validator . Arrays . asList ( staff1 . getEntityId ( ) ) ) ; "<AssertPlaceHolder>" ; } validate ( java . lang . String , org . slc . sli . api . security . context . validator . Set ) { if ( ! ( areParametersValid ( EntityNames . STAFF , entityName , staffIds ) ) ) { return Collections . EMPTY_SET ; } org . slc . sli . api . security . context . validator . Set < java . lang . String > validIds = new org . slc . sli . api . security . context . validator . HashSet < java . lang . String > ( ) ; org . slc . sli . domain . NeutralQuery basicQuery = new org . slc . sli . domain . NeutralQuery ( new org . slc . sli . domain . NeutralCriteria ( "staffReference" , org . slc . sli . domain . NeutralCriteria . CRITERIA_IN , staffIds ) ) ; basicQuery . setIncludeFields ( org . slc . sli . api . security . context . validator . Arrays . asList ( "educationOrganizationReference" , "staffReference" ) ) ; org . slc . sli . api . security . context . validator . TransitiveStaffToStaffValidator . LOG . info ( "Attempting<sp>to<sp>validate<sp>transitively<sp>from<sp>staff<sp>to<sp>staff<sp>with<sp>ids<sp>{}" , staffIds ) ; injectEndDateQuery ( basicQuery ) ; java . lang . Iterable < org . slc . sli . domain . Entity > edOrgAssoc = repo . findAll ( EntityNames . STAFF_ED_ORG_ASSOCIATION , basicQuery ) ; org . slc . sli . api . security . context . validator . Map < java . lang . String , org . slc . sli . api . security . context . validator . Set < java . lang . String > > staffEdorgMap = new org . slc . sli . api . security . context . validator . HashMap < java . lang . String , org . slc . sli . api . security . context . validator . Set < java . lang . String > > ( ) ; populateMapFromMongoResponse ( staffEdorgMap , edOrgAssoc ) ; org . slc . sli . api . security . context . validator . Set < java . lang . String > edOrgLineage = getStaffEdOrgLineage ( ) ; if ( ( edOrgLineage . isEmpty ( ) ) || ( staffEdorgMap . isEmpty ( ) ) ) { return Collections . EMPTY_SET ; } for ( java . util . Map . Entry < java . lang . String , org . slc . sli . api . security . context . validator . Set < java . lang . String > > entry : staffEdorgMap . entrySet ( ) ) { org . slc . sli . api . security . context . validator . Set < java . lang . String > tmpSet = new org . slc . sli . api . security . context . validator . HashSet < java . lang . String > ( entry . getValue ( ) ) ; tmpSet . retainAll ( edOrgLineage ) ; if ( ( tmpSet . size ( ) ) != 0 ) { validIds . add ( entry . getKey ( ) ) ; } } validIds . addAll ( validateThrough ( EntityNames . STAFF_PROGRAM_ASSOCIATION , "programId" ) ) ; validIds . addAll ( validateThrough ( EntityNames . STAFF_COHORT_ASSOCIATION , "cohortId" ) ) ; basicQuery = new org . slc . sli . domain . NeutralQuery ( new org . slc . sli . domain . NeutralCriteria ( "_id" , "in" , edOrgLineage ) ) ; java . lang . Iterable < org . slc . sli . domain . Entity > edorgs = repo . findAll ( EntityNames . EDUCATION_ORGANIZATION , basicQuery ) ; org . slc . sli . api . security . context . validator . List < java . lang . String > programs = new org . slc . sli . api . security . context . validator . ArrayList < java . lang . String > ( ) ; for ( org . slc . sli . domain . Entity e : edorgs ) { java . lang . Object value = e . getBody ( ) . get ( "programReference" ) ; if ( value != null ) { if ( org . slc . sli . api . security . context . validator . List . class . isAssignableFrom ( value . getClass ( ) ) ) { programs . addAll ( ( ( org . slc . sli . api . security . context . validator . List < java . lang . String > ) ( value ) ) ) ; } else if ( java . lang . String . class . isAssignableFrom ( value . getClass ( ) ) ) { programs . add ( ( ( java . lang . String ) ( value ) ) ) ; } } } validIds . addAll ( getIds ( EntityNames . STAFF_PROGRAM_ASSOCIATION , "programId" , programs ) ) ; basicQuery = new org . slc . sli . domain . NeutralQuery ( new org . slc . sli . domain . NeutralCriteria ( "educationOrgId" , "in" , edOrgLineage ) ) ; org . slc . sli . api . security . context . validator . List < java . lang . String > cohorts = ( ( org . slc . sli . api . security . context . validator . List < java . lang . String > ) ( repo . findAllIds ( EntityNames | org . junit . Assert . assertFalse ( validator . validate ( EntityNames . STAFF , ids ) . equals ( ids ) ) |
testReadBoolean ( ) { javax . ws . rs . ext . MessageBodyReader p = new org . apache . cxf . jaxrs . provider . PrimitiveTextProvider ( ) ; boolean valueRead = ( ( java . lang . Boolean ) ( p . readFrom ( Boolean . TYPE , null , null , null , null , new java . io . ByteArrayInputStream ( "true" . getBytes ( ) ) ) ) ) ; "<AssertPlaceHolder>" ; } readFrom ( java . lang . Class , java . lang . reflect . Type , java . lang . annotation . Annotation [ ] , javax . ws . rs . core . MediaType , javax . ws . rs . core . MultivaluedMap , java . io . InputStream ) { return super . readFrom ( type , genericType , anns , mt , headers , is ) ; } | org . junit . Assert . assertTrue ( valueRead ) |
testBasic ( ) { com . arjuna . ats . arjuna . AtomicAction A = new com . arjuna . ats . arjuna . AtomicAction ( ) ; com . hp . mwtests . ts . txoj . common . resources . AtomicObject B = new com . hp . mwtests . ts . txoj . common . resources . AtomicObject ( ) ; A . begin ( ) ; com . arjuna . ats . internal . txoj . abstractrecords . CadaverLockRecord clr = new com . arjuna . ats . internal . txoj . abstractrecords . CadaverLockRecord ( null , B , A ) ; com . arjuna . ats . internal . txoj . abstractrecords . LockRecord lr = new com . arjuna . ats . internal . txoj . abstractrecords . LockRecord ( B , A ) ; "<AssertPlaceHolder>" ; clr . print ( new java . io . PrintWriter ( new java . io . ByteArrayOutputStream ( ) ) ) ; clr . replace ( lr ) ; A . abort ( ) ; } type ( ) { return "/StateManager/AbstractRecord/" + ( getClass ( ) . getName ( ) ) ; } | org . junit . Assert . assertTrue ( ( ( clr . type ( ) ) != null ) ) |
shouldFailWithCorrectPartialResultWhenEscapedKeyIsEmpty ( ) { java . lang . String result = null ; try { processor . process ( "{:}" , withoutContext ( ) , com . github . resource4j . resources . processors . BasicValuePostProcessorTest . RESOLVER ) ; } catch ( com . github . resource4j . resources . processors . ValuePostProcessingException e ) { result = e . getPartialResult ( ) ; } "<AssertPlaceHolder>" ; } getPartialResult ( ) { return partialResult ; } | org . junit . Assert . assertEquals ( "{:}" , result ) |
testSetCustomItemHeight ( ) { list . setData ( RWT . CUSTOM_ITEM_HEIGHT , new java . lang . Integer ( 123 ) ) ; "<AssertPlaceHolder>" ; } getItemHeight ( ) { checkWidget ( ) ; if ( ( customItemHeight ) == ( - 1 ) ) { if ( ! ( layoutCache . hasItemHeight ( ) ) ) { layoutCache . itemHeight = computeItemHeight ( ) ; } return layoutCache . itemHeight ; } return customItemHeight ; } | org . junit . Assert . assertEquals ( 123 , list . getItemHeight ( ) ) |
testPayoutResponseSuccess ( ) { final com . github . kpavlov . restws . server . model . Foo foo = new com . github . kpavlov . restws . server . model . Foo ( "aa" ) ; com . github . kpavlov . restws . server . model . AbstractResponseWrapper < com . github . kpavlov . restws . server . model . Foo > pw = new com . github . kpavlov . restws . server . model . FooResponseWrapper ( ) ; pw . setData ( foo ) ; final java . lang . String str = objectMapper . writeValueAsString ( new com . github . kpavlov . restws . server . model . FooRequestWrapper ( foo ) ) ; final com . github . kpavlov . restws . server . model . Foo foo2 = objectMapper . readValue ( str , com . github . kpavlov . restws . server . model . FooRequestWrapper . class ) . getData ( ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertThat ( foo2 . getName ( ) , org . hamcrest . CoreMatchers . is ( foo . getName ( ) ) ) |
shortCircuitedAvailableInGetFallbackUsingQueue ( ) { setUpTenacityCommand ( 2 , 100 ) ; final com . yammer . tenacity . core . TenacityCommand < ? > exceptionCommand = exceptionCommand ( ) ; exceptionCommand . execute ( ) ; while ( ! ( exceptionCommand . isCircuitBreakerOpen ( ) ) ) ; final java . util . concurrent . Future < java . lang . Boolean > result = shortCircuitedCommand ( ) . queue ( ) ; while ( ! ( result . isDone ( ) ) ) { java . lang . Thread . sleep ( 10 ) ; } "<AssertPlaceHolder>" ; } queue ( ) { return build ( ) . queue ( ) ; } | org . junit . Assert . assertTrue ( result . get ( ) ) |
numProcessRecordsCallsTest ( ) { com . amazonaws . services . kinesis . clientlibrary . lib . worker . KinesisClientLibConfiguration workerConfig = new com . amazonaws . services . kinesis . clientlibrary . lib . worker . KinesisClientLibConfiguration ( leaseTable , streamId , credentials , com . amazonaws . services . dynamodbv2 . streamsadapter . functionals . KinesisParametersTest . KCL_WORKER_ID ) . withMaxRecords ( 10 ) . withInitialPositionInStream ( InitialPositionInStream . TRIM_HORIZON ) . withCallProcessRecordsEvenForEmptyRecordList ( true ) . withIdleTimeBetweenReadsInMillis ( com . amazonaws . services . dynamodbv2 . streamsadapter . functionals . KinesisParametersTest . IDLE_TIME_2S ) ; startKCLWorker ( workerConfig ) ; while ( ( ( ( com . amazonaws . services . dynamodbv2 . streamsadapter . util . TestRecordProcessorFactory ) ( recordProcessorFactory ) ) . getNumRecordsProcessed ( ) ) < 0 ) { com . amazonaws . services . dynamodbv2 . streamsadapter . functionals . KinesisParametersTest . LOG . info ( "Sleep<sp>till<sp>RecordProcessor<sp>is<sp>initialized" ) ; java . lang . Thread . sleep ( com . amazonaws . services . dynamodbv2 . streamsadapter . functionals . THREAD_SLEEP_2S ) ; } java . lang . Thread . sleep ( com . amazonaws . services . dynamodbv2 . streamsadapter . functionals . THREAD_SLEEP_5S ) ; shutDownKCLWorker ( ) ; int numGetRecordsCalls = recordProcessorFactory . getNumProcessRecordsCalls ( ) ; com . amazonaws . services . dynamodbv2 . streamsadapter . functionals . KinesisParametersTest . LOG . info ( ( "Num<sp>getRecords<sp>calls:<sp>" + numGetRecordsCalls ) ) ; "<AssertPlaceHolder>" ; } getNumProcessRecordsCalls ( ) { return processRecordsCallCounter ; } | org . junit . Assert . assertTrue ( ( ( numGetRecordsCalls > 0 ) && ( numGetRecordsCalls <= 3 ) ) ) |
saveAll_Integer_Test ( ) { java . util . List < de . bytefish . pgbulkinsert . pgsql . handlers . PgBulkInsertPrimitivesTest . SampleEntity > entities = new java . util . ArrayList ( ) ; de . bytefish . pgbulkinsert . pgsql . handlers . PgBulkInsertPrimitivesTest . SampleEntity entity = new de . bytefish . pgbulkinsert . pgsql . handlers . PgBulkInsertPrimitivesTest . SampleEntity ( ) ; entity . col_integer = 1 ; entities . add ( entity ) ; de . bytefish . pgbulkinsert . PgBulkInsert < de . bytefish . pgbulkinsert . pgsql . handlers . PgBulkInsertPrimitivesTest . SampleEntity > pgBulkInsert = new de . bytefish . pgbulkinsert . PgBulkInsert ( new de . bytefish . pgbulkinsert . pgsql . handlers . PgBulkInsertPrimitivesTest . SampleEntityMapping ( ) ) ; pgBulkInsert . saveAll ( de . bytefish . pgbulkinsert . util . PostgreSqlUtils . getPGConnection ( connection ) , entities . stream ( ) ) ; java . sql . ResultSet rs = getAll ( ) ; while ( rs . next ( ) ) { int v = rs . getInt ( "col_integer" ) ; "<AssertPlaceHolder>" ; } } getAll ( ) { java . lang . String sqlStatement = java . lang . String . format ( "SELECT<sp>*<sp>FROM<sp>%s.unit_test" , schema ) ; java . sql . Statement statement = connection . createStatement ( ) ; return statement . executeQuery ( sqlStatement ) ; } | org . junit . Assert . assertEquals ( 1 , v ) |
test03DbShouldBeEmpty ( ) { "<AssertPlaceHolder>" ; } countByQuery ( ) { com . orientechnologies . orient . object . db . OObjectDatabaseTx db = dbf . db ( ) ; com . orientechnologies . orient . core . sql . query . OSQLSynchQuery < com . orientechnologies . orient . core . record . impl . ODocument > q = new com . orientechnologies . orient . core . sql . query . OSQLSynchQuery < com . orientechnologies . orient . core . record . impl . ODocument > ( "select<sp>count(*)<sp>from<sp>Person" ) ; java . util . List < com . orientechnologies . orient . core . record . impl . ODocument > results = db . query ( q ) ; assert ( results . size ( ) ) == 1 ; return results . get ( 0 ) . field ( "count" ) ; } | org . junit . Assert . assertThat ( service . countByQuery ( ) , org . hamcrest . CoreMatchers . is ( 0L ) ) |
bnodeTest ( ) { org . apache . jena . arq . querybuilder . Expr e = factory . bnode ( ) ; "<AssertPlaceHolder>" ; } bnode ( ) { return new org . apache . jena . sparql . expr . E_BNode ( ) ; } | org . junit . Assert . assertTrue ( ( e instanceof org . apache . jena . arq . querybuilder . E_BNode ) ) |
testReadInt ( ) { java . io . ByteArrayOutputStream out = new java . io . ByteArrayOutputStream ( ) ; java . io . InputStream in ; try { converter . writeInt ( out , com . ibm . ws . jmx . connector . converter . JSONConverterTest . TEST_INT ) ; in = new java . io . ByteArrayInputStream ( out . toByteArray ( ) ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( ( "Exception<sp>encountered<sp>" + e ) ) ; } } readInt ( java . io . InputStream ) { com . ibm . json . java . JSONArray json = parseArray ( in ) ; if ( ( json . size ( ) ) != 1 ) { com . ibm . ws . jmx . connector . converter . JSONConverter . throwConversionException ( "readInt()<sp>expects<sp>one<sp>item<sp>in<sp>the<sp>array:<sp>[<sp>Integer<sp>]." , json ) ; } return readIntInternal ( json . get ( 0 ) ) ; } | org . junit . Assert . assertEquals ( converter . readInt ( in ) , com . ibm . ws . jmx . connector . converter . JSONConverterTest . TEST_INT ) |
removeNonContainedPart ( ) { org . uberfire . client . workbench . part . WorkbenchPartPresenter mockPresenter = mock ( org . uberfire . client . workbench . part . WorkbenchPartPresenter . class ) ; org . uberfire . client . workbench . part . WorkbenchPartPresenter . View mockPartView = mock ( WorkbenchPartPresenter . View . class ) ; org . uberfire . workbench . model . PartDefinition mockPartDefinition = new org . uberfire . workbench . model . impl . PartDefinitionImpl ( new org . uberfire . mvp . impl . DefaultPlaceRequest ( "mock1" ) ) ; when ( mockPartView . getPresenter ( ) ) . thenReturn ( mockPresenter ) ; when ( mockPresenter . getDefinition ( ) ) . thenReturn ( mockPartDefinition ) ; org . uberfire . client . workbench . part . WorkbenchPartPresenter mockPresenter2 = mock ( org . uberfire . client . workbench . part . WorkbenchPartPresenter . class ) ; org . uberfire . client . workbench . part . WorkbenchPartPresenter . View mockPartView2 = mock ( WorkbenchPartPresenter . View . class ) ; org . uberfire . workbench . model . PartDefinition mockPartDefinition2 = new org . uberfire . workbench . model . impl . PartDefinitionImpl ( new org . uberfire . mvp . impl . DefaultPlaceRequest ( "mock2" ) ) ; when ( mockPartView2 . getPresenter ( ) ) . thenReturn ( mockPresenter2 ) ; when ( mockPresenter2 . getDefinition ( ) ) . thenReturn ( mockPartDefinition2 ) ; when ( view . panel . getPartView ( ) ) . thenReturn ( null ) ; view . addPart ( mockPartView ) ; when ( view . panel . getPartView ( ) ) . thenReturn ( mockPartView ) ; boolean removed = view . removePart ( mockPartDefinition2 ) ; "<AssertPlaceHolder>" ; verify ( panel , never ( ) ) . clear ( ) ; } removePart ( org . uberfire . workbench . model . PartDefinition ) { partTitles . remove ( part ) ; partView . remove ( part ) ; final com . google . gwt . user . client . ui . Widget option = partOptions . remove ( part ) ; this . remove ( option ) ; } | org . junit . Assert . assertFalse ( removed ) |
whenDoesNotContainPeriod ( ) { final java . lang . String actual = org . estatio . dnplugins . EstatioIdentifierFactory . generateIdentifierNameForJavaName ( "LEASETERM_ID" , IdentifierCase . UPPER_CASE , "_" ) ; "<AssertPlaceHolder>" ; } generateIdentifierNameForJavaName ( java . lang . String , org . datanucleus . store . rdbms . identifier . IdentifierCase , java . lang . String ) { if ( javaName == null ) { return null ; } java . lang . StringBuffer s = new java . lang . StringBuffer ( ) ; boolean skipUntilUnderscore = false ; for ( int i = 0 ; i < ( javaName . length ( ) ) ; ++ i ) { char c = javaName . charAt ( i ) ; if ( skipUntilUnderscore ) { if ( c == '_' ) { skipUntilUnderscore = false ; } continue ; } if ( ( ( c >= 'A' ) && ( c <= 'Z' ) ) && ( ( identifierCase != ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE ) ) && ( identifierCase != ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE_QUOTED ) ) ) ) { s . append ( c ) ; } else if ( ( ( c >= 'A' ) && ( c <= 'Z' ) ) && ( ( identifierCase == ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE ) ) || ( identifierCase == ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE_QUOTED ) ) ) ) { s . append ( c ) ; } else if ( ( ( c >= 'a' ) && ( c <= 'z' ) ) && ( ( identifierCase == ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE ) ) || ( identifierCase == ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE_QUOTED ) ) ) ) { s . append ( c ) ; } else if ( ( ( c >= 'a' ) && ( c <= 'z' ) ) && ( ( identifierCase != ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE ) ) && ( identifierCase != ( org . datanucleus . store . rdbms . identifier . IdentifierCase . MIXED_CASE_QUOTED ) ) ) ) { s . append ( ( ( char ) ( c - ( 'a' - 'A' ) ) ) ) ; } else if ( ( ( c >= '0' ) && ( c <= '9' ) ) || ( c == '_' ) ) { s . append ( c ) ; } else if ( c == '.' ) { s . append ( wordSeparator ) ; skipUntilUnderscore = true ; } else { java . lang . String cval = "000" + ( java . lang . Integer . toHexString ( c ) ) ; s . append ( cval . substring ( ( ( cval . length ( ) ) - ( c > 255 ? 4 : 2 ) ) ) ) ; } } while ( ( ( s . length ( ) ) > 0 ) && ( ( s . charAt ( 0 ) ) == '_' ) ) { s . deleteCharAt ( 0 ) ; } if ( ( s . length ( ) ) == 0 ) { throw new java . lang . IllegalArgumentException ( ( "Illegal<sp>Java<sp>identifier:<sp>" + javaName ) ) ; } return s . toString ( ) ; } | org . junit . Assert . assertThat ( actual , org . hamcrest . CoreMatchers . is ( "LEASETERM_ID" ) ) |
isStaticInitializerShouldReturnTrueWhenMethodIsStaticInitializer ( ) { final org . pitest . mutationtest . engine . gregor . MethodInfo testee = this . methodInfo . withMethodName ( "<clinit>" ) ; "<AssertPlaceHolder>" ; } isStaticInitializer ( ) { return "<clinit>" . equals ( this . methodName ) ; } | org . junit . Assert . assertThat ( testee . isStaticInitializer ( ) , org . hamcrest . Matchers . is ( true ) ) |
testListNull ( ) { "<AssertPlaceHolder>" ; } convert ( java . lang . Object ) { T t = null ; if ( attribute != null ) { if ( clazz . isAssignableFrom ( attribute . getClass ( ) ) ) { t = ( ( T ) ( attribute ) ) ; } else if ( attribute instanceof java . util . List ) { final java . util . List l = ( ( java . util . List ) ( attribute ) ) ; if ( ( l . size ( ) ) > 0 ) { final java . lang . Object element = l . get ( 0 ) ; if ( clazz . isAssignableFrom ( element . getClass ( ) ) ) { t = ( ( T ) ( element ) ) ; } } } else { t = internalConvert ( attribute ) ; } } if ( t != null ) { return t ; } else { return defaultValue ( ) ; } } | org . junit . Assert . assertNull ( this . converter . convert ( new java . util . ArrayList ( ) ) ) |
testDynamicQueryByPrimaryKeyMissing ( ) { com . liferay . portal . kernel . dao . orm . DynamicQuery dynamicQuery = com . liferay . portal . kernel . dao . orm . DynamicQueryFactoryUtil . forClass ( com . liferay . trash . model . TrashVersion . class , _dynamicQueryClassLoader ) ; dynamicQuery . add ( com . liferay . portal . kernel . dao . orm . RestrictionsFactoryUtil . eq ( "versionId" , com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ) ) ; java . util . List < com . liferay . trash . model . TrashVersion > result = _persistence . findWithDynamicQuery ( dynamicQuery ) ; "<AssertPlaceHolder>" ; } size ( ) { if ( ( _workflowTaskAssignees ) != null ) { return _workflowTaskAssignees . size ( ) ; } return _kaleoTaskAssignmentInstanceLocalService . getKaleoTaskAssignmentInstancesCount ( _kaleoTaskInstanceToken . getKaleoTaskInstanceTokenId ( ) ) ; } | org . junit . Assert . assertEquals ( 0 , result . size ( ) ) |
testComponentFormatSad ( ) { org . sonatype . nexus . selector . Selector selector = buildSelector ( "component.format<sp>==<sp>'nuget'" ) ; "<AssertPlaceHolder>" ; } evaluate ( org . sonatype . nexus . selector . VariableSource ) { return java . lang . Boolean . TRUE . equals ( expression . evaluate ( org . sonatype . nexus . selector . JexlSelector . asJexlContext ( source ) ) ) ; } | org . junit . Assert . assertFalse ( selector . evaluate ( source ) ) |
testSetParams ( ) { org . deeplearning4j . nn . layers . NeuralNetConfiguration conf = new org . deeplearning4j . nn . layers . NeuralNetConfiguration . Builder ( ) . optimizationAlgo ( OptimizationAlgorithm . LINE_GRADIENT_DESCENT ) . updater ( new org . nd4j . linalg . learning . config . Sgd ( 0.1 ) ) . layer ( new org . deeplearning4j . nn . conf . layers . OutputLayer . Builder ( ) . nIn ( 4 ) . nOut ( 3 ) . weightInit ( WeightInit . ZERO ) . activation ( Activation . SOFTMAX ) . lossFunction ( LossFunctions . LossFunction . MCXENT ) . build ( ) ) . build ( ) ; long numParams = conf . getLayer ( ) . initializer ( ) . numParams ( conf ) ; org . nd4j . linalg . api . ndarray . INDArray params = org . nd4j . linalg . factory . Nd4j . create ( 1 , numParams ) ; org . deeplearning4j . nn . layers . OutputLayer l = ( ( org . deeplearning4j . nn . layers . OutputLayer ) ( conf . getLayer ( ) . instantiate ( conf , java . util . Collections . < org . deeplearning4j . optimize . api . TrainingListener > singletonList ( new org . deeplearning4j . optimize . listeners . ScoreIterationListener ( 1 ) ) , 0 , params , true ) ) ) ; params = l . params ( ) ; l . setParams ( params ) ; "<AssertPlaceHolder>" ; } params ( ) { return null ; } | org . junit . Assert . assertEquals ( params , l . params ( ) ) |
testOriginalNotTcf ( ) { java . io . File targetFolder = testContext . getTestOutputFolder ( ) ; org . apache . uima . jcas . JCas jcas = org . apache . uima . fit . factory . JCasFactory . createJCas ( ) ; de . tudarmstadt . ukp . dkpro . core . api . metadata . type . DocumentMetaData meta = de . tudarmstadt . ukp . dkpro . core . api . metadata . type . DocumentMetaData . create ( jcas ) ; meta . setDocumentBaseUri ( new java . io . File ( "src/test/resources" ) . toURI ( ) . toURL ( ) . toString ( ) ) ; meta . setDocumentUri ( new java . io . File ( "src/test/resources/not-a-tcf-file.txt" ) . toURI ( ) . toURL ( ) . toString ( ) ) ; jcas . setDocumentText ( "okeydokey" ) ; new de . tudarmstadt . ukp . dkpro . core . api . segmentation . type . Token ( jcas , 0 , jcas . getDocumentText ( ) . length ( ) ) . addToIndexes ( ) ; org . apache . uima . analysis_engine . AnalysisEngineDescription writer = createEngineDescription ( de . tudarmstadt . ukp . dkpro . core . io . tcf . TcfWriter . class , TcfWriter . PARAM_TARGET_LOCATION , targetFolder , TcfWriter . PARAM_OVERWRITE , true ) ; org . apache . uima . fit . pipeline . SimplePipeline . runPipeline ( jcas , writer ) ; org . apache . uima . collection . CollectionReaderDescription reader = createReaderDescription ( de . tudarmstadt . ukp . dkpro . core . io . tcf . TcfReader . class , TcfReader . PARAM_SOURCE_LOCATION , ( ( targetFolder . getPath ( ) ) + "/*.tcf" ) ) ; for ( org . apache . uima . jcas . JCas jcas2 : org . apache . uima . fit . pipeline . SimplePipeline . iteratePipeline ( reader ) ) { "<AssertPlaceHolder>" ; } } getPath ( ) { return path ; } | org . junit . Assert . assertEquals ( "okeydokey" , jcas2 . getDocumentText ( ) ) |
checkUpdate ( ) { org . eclipse . ice . datastructures . entry . StringEntry testKey = new org . eclipse . ice . datastructures . entry . StringEntry ( ) ; org . eclipse . ice . datastructures . entry . StringEntry testValue = new org . eclipse . ice . datastructures . entry . StringEntry ( ) ; org . eclipse . ice . vibe . kvPair . VibeKVPairRow row = new org . eclipse . ice . vibe . kvPair . VibeKVPairRow ( testKey , testValue ) ; org . eclipse . ice . tests . vibe . kvPair . VibeKVPairRowTester . TestListener listener = new org . eclipse . ice . tests . vibe . kvPair . VibeKVPairRowTester . TestListener ( row ) ; listener . wasUpdated ( ) ; testValue . setValue ( "test" ) ; try { java . lang . Thread . sleep ( 100 ) ; } catch ( java . lang . InterruptedException e ) { e . printStackTrace ( ) ; org . junit . Assert . fail ( "Exception<sp>while<sp>waiting<sp>for<sp>update." ) ; } "<AssertPlaceHolder>" ; } wasUpdated ( ) { return updated ; } | org . junit . Assert . assertTrue ( listener . wasUpdated ( ) ) |
_ ( ) { java . lang . String targetFileName = java . lang . String . format ( "adsWriteFailure_%s.log.%d" , "000" , 12345678901234L ) ; "<AssertPlaceHolder>" ; } checkAcceptResult ( java . lang . String ) { java . io . File file = new java . io . File ( targetFileName ) ; file . deleteOnExit ( ) ; try { file . createNewFile ( ) ; com . fujitsu . dc . common . ads . AdsWriteFailureLogFilter filter = new com . fujitsu . dc . common . ads . AdsWriteFailureLogFilter ( "000" ) ; return filter . accept ( new java . io . File ( "." ) , targetFileName ) ; } finally { file . delete ( ) ; } } | org . junit . Assert . assertFalse ( checkAcceptResult ( targetFileName ) ) |
shouldReturnTrueWhenEqualToStartTime ( ) { java . util . Calendar cal = org . openstack . atlas . util . common . CalendarUtils . stringToCalendar ( "2013-01-29<sp>00:00:00" ) ; java . util . Calendar startTime = org . openstack . atlas . util . common . CalendarUtils . stringToCalendar ( "2013-01-29<sp>00:00:00" ) ; java . util . Calendar endTime = org . openstack . atlas . util . common . CalendarUtils . stringToCalendar ( "2013-01-29<sp>01:00:00" ) ; boolean isBetween = org . openstack . atlas . util . common . CalendarUtils . isBetween ( cal , startTime , endTime , true ) ; "<AssertPlaceHolder>" ; } isBetween ( java . util . Calendar , java . util . Calendar , java . util . Calendar , boolean ) { if ( endTimeInclusive ) { return ( ( cal . compareTo ( startTime ) ) >= 0 ) && ( ( cal . compareTo ( endTime ) ) <= 0 ) ; } return ( ( cal . compareTo ( startTime ) ) >= 0 ) && ( ( cal . compareTo ( endTime ) ) < 0 ) ; } | org . junit . Assert . assertTrue ( isBetween ) |
clearAll ( ) { org . apache . flink . runtime . io . network . partition . ResultPartitionID partitionId = new org . apache . flink . runtime . io . network . partition . ResultPartitionID ( ) ; org . apache . flink . runtime . io . network . TaskEventDispatcher ted = new org . apache . flink . runtime . io . network . TaskEventDispatcher ( ) ; ted . registerPartition ( partitionId ) ; org . apache . flink . runtime . io . network . TaskEventDispatcherTest . ZeroShotEventListener eventListener1 = new org . apache . flink . runtime . io . network . TaskEventDispatcherTest . ZeroShotEventListener ( ) ; ted . subscribeToEvent ( partitionId , eventListener1 , org . apache . flink . runtime . iterative . event . AllWorkersDoneEvent . class ) ; ted . clearAll ( ) ; "<AssertPlaceHolder>" ; } publish ( org . apache . flink . runtime . io . network . partition . ResultPartitionID , org . apache . flink . runtime . event . TaskEvent ) { checkNotNull ( partitionId ) ; checkNotNull ( event ) ; org . apache . flink . runtime . io . network . api . TaskEventHandler taskEventHandler ; synchronized ( registeredHandlers ) { taskEventHandler = registeredHandlers . get ( partitionId ) ; } if ( taskEventHandler != null ) { taskEventHandler . publish ( event ) ; return true ; } return false ; } | org . junit . Assert . assertFalse ( ted . publish ( partitionId , new org . apache . flink . runtime . iterative . event . AllWorkersDoneEvent ( ) ) ) |
testEqualQueryStatsWithLastInvocation ( ) { java . util . Comparator < org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport . QueryStats > queryStatsComparator = createComparator ( ) ; org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport . QueryStats q1 = new org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport . QueryStats ( "abc" ) ; q1 . add ( 0 , 100 ) ; "<AssertPlaceHolder>" ; } compare ( org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport$QueryStats , org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport$QueryStats ) { return java . lang . Long . compare ( org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport . QueryStatsComparator . handleZero ( stats1 . lastInvocation ) , org . apache . tomcat . jdbc . pool . interceptor . SlowQueryReport . QueryStatsComparator . handleZero ( stats2 . lastInvocation ) ) ; } | org . junit . Assert . assertEquals ( 0 , queryStatsComparator . compare ( q1 , q1 ) ) |
dependencyMaintainedForSet ( ) { org . teavm . classlib . java . lang . reflect . FieldTest . ReflectableType instance = new org . teavm . classlib . java . lang . reflect . FieldTest . ReflectableType ( ) ; java . lang . reflect . Field field = org . teavm . classlib . java . lang . reflect . FieldTest . ReflectableType . class . getDeclaredField ( "c" ) ; field . set ( instance , new org . teavm . classlib . java . lang . reflect . FieldTest . Foo ( 123 ) ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return value ; } | org . junit . Assert . assertEquals ( 123 , ( ( org . teavm . classlib . java . lang . reflect . FieldTest . Foo ) ( instance . c ) ) . getValue ( ) ) |
resizerWhereSourceLargerThanDestination_MoreThan2Times ( ) { net . coobird . thumbnailator . resizers . Resizer resizer = Resizers . PROGRESSIVE ; net . coobird . thumbnailator . resizers . ResizerFactory factory = net . coobird . thumbnailator . resizers . DefaultResizerFactory . getInstance ( ) ; java . awt . Dimension sourceDimension = new java . awt . Dimension ( 200 , 200 ) ; java . awt . Dimension targetDimension = new java . awt . Dimension ( 50 , 50 ) ; net . coobird . thumbnailator . resizers . Resizer receivedResizer = factory . getResizer ( sourceDimension , targetDimension ) ; "<AssertPlaceHolder>" ; } getResizer ( java . awt . Dimension , java . awt . Dimension ) { return resizer ; } | org . junit . Assert . assertEquals ( resizer , receivedResizer ) |
testSuperstepOne ( ) { org . apache . giraph . graph . Vertex < org . apache . hadoop . io . IntWritable , org . apache . giraph . examples . SimpleTriangleClosingComputation . IntArrayListWritable , org . apache . hadoop . io . NullWritable > vertex = new org . apache . giraph . graph . DefaultVertex < org . apache . hadoop . io . IntWritable , org . apache . giraph . examples . SimpleTriangleClosingComputation . IntArrayListWritable , org . apache . hadoop . io . NullWritable > ( ) ; org . apache . giraph . examples . SimpleTriangleClosingComputation computation = new org . apache . giraph . examples . SimpleTriangleClosingComputation ( ) ; org . apache . giraph . utils . MockUtils . MockedEnvironment env = org . apache . giraph . utils . MockUtils . prepareVertexAndComputation ( vertex , new org . apache . hadoop . io . IntWritable ( 1 ) , null , false , computation , 1L ) ; computation . compute ( vertex , com . google . common . collect . Lists . < org . apache . hadoop . io . IntWritable > newArrayList ( new org . apache . hadoop . io . IntWritable ( 7 ) , new org . apache . hadoop . io . IntWritable ( 3 ) , new org . apache . hadoop . io . IntWritable ( 4 ) , new org . apache . hadoop . io . IntWritable ( 7 ) , new org . apache . hadoop . io . IntWritable ( 4 ) , new org . apache . hadoop . io . IntWritable ( 2 ) , new org . apache . hadoop . io . IntWritable ( 4 ) ) ) ; final java . lang . String pairCheck = "[4,<sp>7]" ; "<AssertPlaceHolder>" ; } getValue ( ) { return value ; } | org . junit . Assert . assertEquals ( pairCheck , vertex . getValue ( ) . toString ( ) ) |
testGetUnpackSizeOne ( ) { org . apache . commons . compress . archivers . sevenz . Folder folder = new org . apache . commons . compress . archivers . sevenz . Folder ( ) ; folder . totalOutputStreams = 266L ; org . apache . commons . compress . archivers . sevenz . BindPair [ ] bindPairArray = new org . apache . commons . compress . archivers . sevenz . BindPair [ 1 ] ; org . apache . commons . compress . archivers . sevenz . BindPair bindPair = new org . apache . commons . compress . archivers . sevenz . BindPair ( ) ; bindPairArray [ 0 ] = bindPair ; folder . bindPairs = bindPairArray ; folder . totalOutputStreams = 1L ; "<AssertPlaceHolder>" ; } getUnpackSize ( ) { if ( ( totalOutputStreams ) == 0 ) { return 0 ; } for ( int i = ( ( int ) ( totalOutputStreams ) ) - 1 ; i >= 0 ; i -- ) { if ( ( findBindPairForOutStream ( i ) ) < 0 ) { return unpackSizes [ i ] ; } } return 0 ; } | org . junit . Assert . assertEquals ( 0L , folder . getUnpackSize ( ) ) |
parse_noparens ( ) { com . psddev . dari . db . Predicate pred = parser . parse ( "a<sp>=<sp>1<sp>and<sp>b<sp>=<sp>2<sp>or<sp>c<sp>=<sp>3" ) ; com . psddev . dari . db . Predicate expect = com . psddev . dari . db . CompoundPredicate . combine ( PredicateParser . OR_OPERATOR , com . psddev . dari . db . CompoundPredicate . combine ( PredicateParser . AND_OPERATOR , new com . psddev . dari . db . ComparisonPredicate ( PredicateParser . EQUALS_ANY_OPERATOR , false , "a" , java . util . Arrays . asList ( "1" ) ) , new com . psddev . dari . db . ComparisonPredicate ( PredicateParser . EQUALS_ANY_OPERATOR , false , "b" , java . util . Arrays . asList ( "2" ) ) ) , new com . psddev . dari . db . ComparisonPredicate ( PredicateParser . EQUALS_ANY_OPERATOR , false , "c" , java . util . Arrays . asList ( "3" ) ) ) ; "<AssertPlaceHolder>" ; } parse ( byte [ ] ) { try { return parseAny ( bytes ) ; } catch ( org . codehaus . jackson . JsonParseException error ) { throw new com . psddev . dari . util . JsonParsingException ( "Can't<sp>parse<sp>JSON<sp>bytes!" , error ) ; } catch ( java . io . IOException error ) { throw new java . lang . IllegalStateException ( error ) ; } } | org . junit . Assert . assertEquals ( expect , pred ) |
testDatumEindeNaDatumAanvang ( ) { final nl . bzk . brp . model . hisvolledig . momentview . kern . HuwelijkView huidigeSituatie = maakbestaandeSituatie ( 20101010 ) ; final nl . bzk . brp . model . bericht . kern . HuwelijkBericht nieuweSituatie = maakNieuweSituatie ( 20101011 ) ; final java . util . List < nl . bzk . brp . model . basis . BerichtEntiteit > berichtEntiteits = bral2113 . voerRegelUit ( huidigeSituatie , nieuweSituatie , null , null ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , berichtEntiteits . size ( ) ) |
testRevWalkDisposeClosesReader ( ) { try ( org . eclipse . jgit . lib . Repository repo = com . baeldung . jgit . helper . Helper . openJGitRepository ( ) ) { try ( org . eclipse . jgit . lib . ObjectReader reader = repo . newObjectReader ( ) ) { try ( org . eclipse . jgit . revwalk . RevWalk walk = new org . eclipse . jgit . revwalk . RevWalk ( reader ) ) { walk . dispose ( ) ; org . eclipse . jgit . lib . Ref head = repo . exactRef ( "refs/heads/master" ) ; System . out . println ( ( "Found<sp>head:<sp>" + head ) ) ; org . eclipse . jgit . lib . ObjectLoader loader = reader . open ( head . getObjectId ( ) ) ; "<AssertPlaceHolder>" ; } } } } open ( java . io . Serializable ) { tokens = new java . util . StringTokenizer ( "1,2,3,4,5,6,7,8,9,10" , "," ) ; if ( checkpoint != null ) { while ( ! ( java . lang . Integer . valueOf ( tokens . nextToken ( ) ) . equals ( ( ( com . baeldung . batch . understanding . exception . MyInputRecord ) ( checkpoint ) ) . getId ( ) ) ) ) { } } } | org . junit . Assert . assertNotNull ( loader ) |
shouldDoVerifyRegularTextNotPresent ( ) { given ( commandProcessor . doCommand ( "isTextPresent" , new java . lang . String [ ] { "foo" } ) ) . willReturn ( "false" ) ; final boolean result = seleniumDriverFixture . doOn ( "verifyTextNotPresent" , "foo" ) ; "<AssertPlaceHolder>" ; } is ( java . lang . String ) { com . xebia . incubator . xebium . SeleniumDriverFixture . LOG . info ( ( ( "Obtain<sp>result<sp>from<sp>|<sp>" + command ) + "<sp>|" ) ) ; return is ( command , new java . lang . String [ ] { } ) ; } | org . junit . Assert . assertThat ( result , org . hamcrest . Matchers . is ( true ) ) |
testGetPrincipalsLocalGroup ( ) { java . util . Set < ? extends java . security . Principal > principals = principalProvider . getPrincipals ( createTestGroup ( ) . getID ( ) ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( size ) == 0 ; } | org . junit . Assert . assertTrue ( principals . isEmpty ( ) ) |
getValue ( ) { mandatorySubject . setValue ( "test" ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return value ; } | org . junit . Assert . assertEquals ( "test" , mandatorySubject . getValue ( ) ) |
difference_of_sets_java ( ) { java . util . Set < java . lang . String > possibleFriendRequests = new java . util . HashSet < java . lang . String > ( yourFriends ) ; possibleFriendRequests . removeAll ( myFriends ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( 6 , possibleFriendRequests . size ( ) ) |
testHaalAdresEnVulAanMetHistorieNull ( ) { nl . bzk . brp . model . operationeel . kern . PersoonAdresModel nieuw = persoonAdresRepository . vulaanAdresMetHistorie ( null , true ) ; "<AssertPlaceHolder>" ; } vulaanAdresMetHistorie ( nl . bzk . brp . model . operationeel . kern . PersoonModel , boolean ) { if ( ( persoon . getAdressen ( ) ) != null ) { for ( nl . bzk . brp . model . operationeel . kern . PersoonAdresModel adres : persoon . getAdressen ( ) ) { persoonAdresRepository . vulaanAdresMetHistorie ( adres , inclFormeleHistorie ) ; } } } | org . junit . Assert . assertEquals ( null , nieuw ) |
shouldReturnFalseForNullVersion ( ) { org . togglz . appengine . activation . ApplicationVersionActivationStrategy strategy = new org . togglz . appengine . activation . ApplicationVersionActivationStrategy ( ) ; org . togglz . core . repository . FeatureState state = aVersionState ( null ) ; boolean active = strategy . isActive ( state , aFeatureUser ( org . togglz . appengine . activation . ApplicationVersionActivationStrategyTest . JOHN ) ) ; "<AssertPlaceHolder>" ; } aFeatureUser ( java . lang . String ) { return new org . togglz . core . user . SimpleFeatureUser ( string ) ; } | org . junit . Assert . assertFalse ( active ) |
testExportOneCollectionWQuery ( ) { java . io . File f = null ; int entitiesToCreate = 5 ; try { f = new java . io . File ( "exportOneCollectionWQuery.json" ) ; } catch ( java . lang . Exception e ) { } f . deleteOnExit ( ) ; org . apache . usergrid . persistence . EntityManager em = org . apache . usergrid . management . export . ExportServiceIT . setup . getEmf ( ) . getEntityManager ( applicationId ) ; em . createApplicationCollection ( "applicationId" 1 ) ; org . apache . usergrid . management . export . ExportServiceIT . setup . getEntityIndex ( ) . refresh ( applicationId ) ; org . apache . usergrid . management . export . Map < java . lang . String , java . lang . Object > userProperties = null ; org . apache . usergrid . persistence . Entity [ ] entity ; entity = new org . apache . usergrid . persistence . Entity [ entitiesToCreate ] ; for ( int i = 0 ; i < entitiesToCreate ; i ++ ) { userProperties = new org . apache . usergrid . management . export . LinkedHashMap < java . lang . String , java . lang . Object > ( ) ; userProperties . put ( "username" , ( "applicationId" 0 + i ) ) ; userProperties . put ( "email" , ( ( "test" + i ) + "@anuff.com" ) ) ; entity [ i ] = em . create ( "applicationId" 1 , userProperties ) ; } org . apache . usergrid . management . export . S3Export s3Export = new org . apache . usergrid . management . export . MockS3ExportImpl ( "exportOneCollectionWQuery.json" ) ; org . apache . usergrid . management . export . ExportService exportService = org . apache . usergrid . management . export . ExportServiceIT . setup . getExportService ( ) ; java . lang . String appName = newOrgAppAdminRule . getApplicationInfo ( ) . getName ( ) ; org . apache . usergrid . management . export . HashMap < java . lang . String , java . lang . Object > payload = payloadBuilder ( appName ) ; payload . put ( "applicationId" 2 , "select<sp>*<sp>where<sp>username<sp>contains<sp>'billybob0'" ) ; payload . put ( "organizationId" , organization . getUuid ( ) ) ; payload . put ( "applicationId" , applicationId ) ; payload . put ( "collectionName" , "applicationId" 1 ) ; org . apache . usergrid . management . export . UUID exportUUID = exportService . schedule ( payload ) ; org . apache . usergrid . persistence . entities . JobData jobData = jobDataCreator ( payload , exportUUID , s3Export ) ; org . apache . usergrid . batch . JobExecution jobExecution = mock ( org . apache . usergrid . batch . JobExecution . class ) ; when ( jobExecution . getJobData ( ) ) . thenReturn ( jobData ) ; org . apache . usergrid . management . export . ExportServiceIT . setup . getEntityIndex ( ) . refresh ( applicationId ) ; exportService . doExport ( jobExecution ) ; com . fasterxml . jackson . core . type . TypeReference < org . apache . usergrid . management . export . HashMap < java . lang . String , java . lang . Object > > typeRef = new com . fasterxml . jackson . core . type . TypeReference < org . apache . usergrid . management . export . HashMap < java . lang . String , java . lang . Object > > ( ) { } ; com . fasterxml . jackson . databind . ObjectMapper mapper = new com . fasterxml . jackson . databind . ObjectMapper ( ) ; org . apache . usergrid . management . export . Map < java . lang . String , java . lang . Object > jsonMap = mapper . readValue ( new java . io . FileReader ( f ) , typeRef ) ; org . apache . usergrid . management . export . Map collectionsMap = ( ( org . apache . usergrid . management . export . Map ) ( jsonMap . get ( "collections" ) ) ) ; java . lang . String collectionName = ( ( java . lang . String ) ( collectionsMap . keySet ( ) . iterator ( ) . next ( ) ) ) ; org . apache . usergrid . management . export . List collectionList = ( ( org . apache . usergrid . management . export . List ) ( collectionsMap . get ( collectionName ) ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return getDynamicProperties ( ) . size ( ) ; } | org . junit . Assert . assertEquals ( 1 , collectionList . size ( ) ) |
testAddFileAsStream ( ) { org . digidoc4j . Container container = this . createEmptyContainerBy ( Container . DocumentType . BDOC ) ; java . io . ByteArrayInputStream stream = new java . io . ByteArrayInputStream ( "tere,<sp>tere" . getBytes ( ) ) ; container . addDataFile ( stream , "test1.txt" , "text/plain" ) ; this . createSignatureBy ( container , this . pkcs12SignatureToken ) ; java . lang . String file = this . getFileBy ( "bdoc" ) ; container . save ( file ) ; org . digidoc4j . Container containerToTest = org . digidoc4j . ContainerOpener . open ( file ) ; "<AssertPlaceHolder>" ; } getDataFiles ( ) { return m_dataFiles ; } | org . junit . Assert . assertEquals ( "test1.txt" , containerToTest . getDataFiles ( ) . get ( 0 ) . getName ( ) ) |
shouldConvertToUrlEncode ( ) { java . lang . String expectedUrlEncoded = "key1=value1&key2=value2" ; java . lang . String urlEncoded = requestMap . toUrlEncode ( "UTF-8" ) ; "<AssertPlaceHolder>" ; } toUrlEncode ( java . lang . String ) { java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; for ( Map . Entry < java . lang . String , java . lang . String > entry : map . entrySet ( ) ) { if ( ( sb . length ( ) ) > 0 ) { sb . append ( "&" ) ; } sb . append ( java . lang . String . format ( "%s=%s" , java . net . URLEncoder . encode ( entry . getKey ( ) , charset ) , java . net . URLEncoder . encode ( entry . getValue ( ) , charset ) ) ) ; } return sb . toString ( ) ; } | org . junit . Assert . assertEquals ( expectedUrlEncoded , urlEncoded ) |
testDeleteRowAndUpsertValueAtSameTS1 ( ) { try { final org . apache . phoenix . end2end . ConcurrentMutationsIT . MyClock clock = new org . apache . phoenix . end2end . ConcurrentMutationsIT . MyClock ( 1000 ) ; org . apache . phoenix . util . EnvironmentEdgeManager . injectEdge ( clock ) ; java . lang . String tableName = generateUniqueName ( ) ; java . lang . String indexName = generateUniqueName ( ) ; java . util . Properties props = org . apache . phoenix . util . PropertiesUtil . deepCopy ( TestUtil . TEST_PROPERTIES ) ; long ts = 1000 ; clock . time = ts ; java . sql . Connection conn = java . sql . DriverManager . getConnection ( getUrl ( ) , props ) ; conn . createStatement ( ) . execute ( ( ( "CREATE<sp>TABLE<sp>" + tableName ) + "(k1<sp>CHAR(2)<sp>NOT<sp>NULL,<sp>k2<sp>CHAR(2)<sp>NOT<sp>NULL,<sp>ts<sp>TIMESTAMP,<sp>A.V<sp>VARCHAR,<sp>B.V2<sp>VARCHAR,<sp>CONSTRAINT<sp>pk<sp>PRIMARY<sp>KEY<sp>(k1,k2))<sp>COLUMN_ENCODED_BYTES<sp>=<sp>0,<sp>STORE_NULLS=true" ) ) ; conn . close ( ) ; conn = java . sql . DriverManager . getConnection ( getUrl ( ) , props ) ; conn . createStatement ( ) . execute ( ( ( ( ( "CREATE<sp>INDEX<sp>" + indexName ) + "<sp>ON<sp>" ) + tableName ) + "(k2,k1,ts)<sp>INCLUDE<sp>(V,<sp>v2)" ) ) ; conn . close ( ) ; conn = java . sql . DriverManager . getConnection ( getUrl ( ) , props ) ; java . sql . PreparedStatement stmt = conn . prepareStatement ( ( ( "UPSERT<sp>INTO<sp>" + tableName ) + "<sp>VALUES('aa','aa',?,<sp>'0','1')" ) ) ; stmt . setTimestamp ( 1 , new java . sql . Timestamp ( 1000L ) ) ; stmt . executeUpdate ( ) ; conn . commit ( ) ; conn . close ( ) ; java . sql . Timestamp expectedTimestamp ; clock . setAdvance ( false ) ; conn = java . sql . DriverManager . getConnection ( getUrl ( ) , props ) ; stmt = conn . prepareStatement ( ( ( "DELETE<sp>FROM<sp>" + tableName ) + "<sp>WHERE<sp>(K1,K2)<sp>=<sp>('aa','aa')" ) ) ; stmt . executeUpdate ( ) ; conn . commit ( ) ; expectedTimestamp = new java . sql . Timestamp ( 6000L ) ; stmt = conn . prepareStatement ( ( ( "UPSERT<sp>INTO<sp>" + tableName ) + "<sp>VALUES('aa','aa',?,<sp>null,'3')" ) ) ; stmt . setTimestamp ( 1 , expectedTimestamp ) ; stmt . executeUpdate ( ) ; conn . commit ( ) ; clock . setAdvance ( true ) ; conn . close ( ) ; conn = java . sql . DriverManager . getConnection ( getUrl ( ) , props ) ; long rowCount = org . apache . phoenix . util . IndexScrutiny . scrutinizeIndex ( conn , tableName , indexName ) ; "<AssertPlaceHolder>" ; conn . close ( ) ; } finally { org . apache . phoenix . util . EnvironmentEdgeManager . injectEdge ( null ) ; } } | org . junit . Assert . assertEquals ( 0 , rowCount ) |
testMoreThanHalfEle ( ) { java . lang . Integer [ ] data = new java . lang . Integer [ ] { 1 , 2 , 3 , 2 , 2 , 2 , 5 , 4 , 2 } ; "<AssertPlaceHolder>" ; } moreThanHalfEle ( T [ ] ) { com . google . common . base . Preconditions . checkNotNull ( arr ) ; com . google . common . base . Preconditions . checkArgument ( ( ( arr . length ) != 0 ) ) ; T most_ele = null ; int times = 0 ; for ( T ele : arr ) { if ( java . util . Objects . equals ( ele , most_ele ) ) { times ++ ; } else if ( times == 0 ) { times = 1 ; most_ele = ele ; } else { times -- ; } } return ssj . algorithm . ArrayUtil . checkMoreThanHalf ( arr , most_ele ) ? most_ele : null ; } | org . junit . Assert . assertEquals ( java . lang . Integer . valueOf ( 2 ) , ssj . algorithm . ArrayUtil . moreThanHalfEle ( data ) ) |
testBogusProtocol1 ( ) { java . lang . String SERVER_URL = "foo://localhost:31000" ; "<AssertPlaceHolder>" ; } isValidServerURL ( java . lang . String ) { boolean valid = true ; try { new org . teiid . net . TeiidURL ( serverURL ) ; } catch ( java . lang . Exception e ) { valid = false ; } return valid ; } | org . junit . Assert . assertFalse ( org . teiid . net . TeiidURL . isValidServerURL ( SERVER_URL ) ) |
testGatk ( ) { java . lang . String headerLines = GATK_HEADER ; java . lang . String vcfLine = "1\t17452\t.\tG\tA\t35.74\t.\t" + ( ( "AC=2;AF=1.00;AN=2;DP=250;Dels=0.00;FS=0.000;GC=60.85;HRun=0;HaplotypeScore=0.0000;MLEAC=2;MLEAF=1.00;MQ=2.41;" + "MQ0=248;QD=0.14;SOR=2.303\t" ) + "GT:AD:DP:GQ:PL\t1/1:117,133:250:6:63,6,0\n" ) ; htsjdk . variant . variantcontext . VariantContext variant = writeAndReadVcfLine ( vcfLine , headerLines ) ; htsjdk . variant . variantcontext . Genotype gt = variant . getGenotype ( "individual" ) ; "<AssertPlaceHolder>" ; } guessFromGenotype ( htsjdk . variant . variantcontext . Genotype ) { if ( ( ( gt . hasAnyAttribute ( "NR" 1 ) ) && ( gt . hasAnyAttribute ( "DV" ) ) ) && ( gt . hasAnyAttribute ( "DPR" ) ) ) { return de . charite . compbio . jannovar . filter . impl . gt . SupportedVarCaller . BCFTOOLS ; } else if ( ( ( ( ( ( gt . hasAnyAttribute ( "GT" ) ) && ( gt . hasAnyAttribute ( "GQ" ) ) ) && ( gt . hasAnyAttribute ( "RO" ) ) ) && ( gt . hasAnyAttribute ( "NR" 2 ) ) ) && ( gt . hasAnyAttribute ( "AO" ) ) ) && ( gt . hasAnyAttribute ( "QA" ) ) ) { return de . charite . compbio . jannovar . filter . impl . gt . SupportedVarCaller . FREEBAYES ; } else if ( ( ( ( ( gt . hasAnyAttribute ( "GT" ) ) && ( gt . hasAnyAttribute ( "AD" ) ) ) && ( gt . hasAnyAttribute ( "NR" 1 ) ) ) && ( gt . hasAnyAttribute ( "GQ" ) ) ) && ( gt . hasAnyAttribute ( "PL" ) ) ) { return de . charite . compbio . jannovar . filter . impl . gt . SupportedVarCaller . GATK_CALLER ; } else if ( ( ( ( gt . hasAnyAttribute ( "GT" ) ) && ( gt . hasAnyAttribute ( "GQ" ) ) ) && ( gt . hasAnyAttribute ( "NR" ) ) ) && ( gt . hasAnyAttribute ( "NR" 0 ) ) ) { return de . charite . compbio . jannovar . filter . impl . gt . SupportedVarCaller . PLATYPUS ; } else { return de . charite . compbio . jannovar . filter . impl . gt . SupportedVarCaller . GATK_CALLER ; } } | org . junit . Assert . assertEquals ( SupportedVarCaller . GATK_CALLER , de . charite . compbio . jannovar . filter . impl . gt . SupportedVarCaller . guessFromGenotype ( gt ) ) |
returnsEmptyIteratorIfAskedToIterateFromStart ( ) { final java . lang . String stream = generateStreamName ( ) ; eventstore . appendToStream ( stream , ExpectedVersion . NO_STREAM , newTestEvents ( 20 ) ) . join ( ) ; java . util . Iterator < com . github . msemys . esjc . ResolvedEvent > iterator = eventstore . iterateAllEventsBackward ( Position . START , 1 , false ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { return iterator ( ) . hasNext ( ) ; } | org . junit . Assert . assertFalse ( iterator . hasNext ( ) ) |
deveObterCpfComoFoiSetado ( ) { final com . fincatto . documentofiscal . nfe310 . classes . cadastro . NFRetornoConsultaCadastroDados retornoConsultaCadastroDados = new com . fincatto . documentofiscal . nfe310 . classes . cadastro . NFRetornoConsultaCadastroDados ( ) ; retornoConsultaCadastroDados . setCpf ( "24868707841" ) ; "<AssertPlaceHolder>" ; } getCpf ( ) { return this . cpf ; } | org . junit . Assert . assertEquals ( "24868707841" , retornoConsultaCadastroDados . getCpf ( ) ) |
testResourcesUnderLibrary ( ) { com . facebook . buck . core . model . BuildTarget fileTarget = com . facebook . buck . core . model . BuildTargetFactory . newInstance ( rootPath , "//foo" , "file" ) ; com . facebook . buck . core . model . BuildTarget resourceTarget = com . facebook . buck . core . model . BuildTargetFactory . newInstance ( rootPath , "//foo" , "res" ) ; com . facebook . buck . core . model . BuildTarget libraryTarget = com . facebook . buck . core . model . BuildTargetFactory . newInstance ( rootPath , "//foo" , "lib" ) ; com . facebook . buck . core . model . targetgraph . TargetNode < ? > fileNode = new com . facebook . buck . shell . ExportFileBuilder ( fileTarget ) . build ( ) ; com . facebook . buck . core . model . targetgraph . TargetNode < ? > resourceNode = com . facebook . buck . apple . AppleResourceBuilder . createBuilder ( resourceTarget ) . setDirs ( com . google . common . collect . ImmutableSet . of ( ) ) . setFiles ( com . google . common . collect . ImmutableSet . of ( com . facebook . buck . core . sourcepath . DefaultBuildTargetSourcePath . of ( fileTarget ) ) ) . build ( ) ; com . facebook . buck . core . model . targetgraph . TargetNode < ? > libraryNode = com . facebook . buck . apple . AppleLibraryBuilder . createBuilder ( libraryTarget ) . setDeps ( com . google . common . collect . ImmutableSortedSet . of ( resourceTarget ) ) . build ( ) ; com . facebook . buck . features . apple . project . ProjectGenerator projectGenerator = createProjectGenerator ( com . google . common . collect . ImmutableSet . of ( fileNode , resourceNode , libraryNode ) ) ; projectGenerator . createXcodeProjects ( ) ; com . facebook . buck . apple . xcode . xcodeproj . PBXProject project = projectGenerator . getGeneratedProject ( ) ; com . facebook . buck . apple . xcode . xcodeproj . PBXGroup mainGroup = project . getMainGroup ( ) ; com . facebook . buck . apple . xcode . xcodeproj . PBXGroup resourcesGroup = mainGroup . getOrCreateDescendantGroupByPath ( com . google . common . collect . ImmutableList . of ( "//foo:lib" , "Resources" ) ) ; com . facebook . buck . apple . xcode . xcodeproj . PBXFileReference resource = ( ( com . facebook . buck . apple . xcode . xcodeproj . PBXFileReference ) ( com . google . common . collect . Iterables . get ( resourcesGroup . getChildren ( ) , 0 ) ) ) ; "<AssertPlaceHolder>" ; } getName ( ) { return mName ; } | org . junit . Assert . assertThat ( resource . getName ( ) , org . hamcrest . CoreMatchers . equalTo ( "file" ) ) |
testToString ( ) { buffer . clear ( ) ; buffer . writeBytes ( copiedBuffer ( "Hello,<sp>World!" , CharsetUtil . ISO_8859_1 ) ) ; "<AssertPlaceHolder>" ; } toString ( java . nio . charset . Charset ) { return toString ( readerIndex , readableBytes ( ) , charset ) ; } | org . junit . Assert . assertEquals ( "Hello,<sp>World!" , buffer . toString ( CharsetUtil . ISO_8859_1 ) ) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.