input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testFetchByPrimaryKeysWithMultiplePrimaryKeysWhereNoPrimaryKeysExist ( ) { long pk1 = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; long pk2 = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; primaryKeys . add ( pk1 ) ; primaryKeys . add ( pk2 ) ; java . util . Map < java . io . Serializable , com . liferay . layout . page . template . model . LayoutPageTemplateStructureRel > layoutPageTemplateStructureRels = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( layoutPageTemplateStructureRels . isEmpty ( ) )
testGetTableCellRendererComponent ( ) { System . out . println ( "getTableCellRendererComponent" ) ; javax . swing . JTable table = new kg . apc . jmeter . graphs . ChartRowsTable ( null ) ; java . lang . Object value = new java . lang . String ( ) ; boolean isSelected = false ; boolean hasFocus = false ; int row = 0 ; int column = 0 ; kg . apc . jmeter . graphs . HeaderAsTextRenderer instance = new kg . apc . jmeter . graphs . HeaderAsTextRenderer ( ) ; java . awt . Component result = instance . getTableCellRendererComponent ( table , value , isSelected , hasFocus , row , column ) ; "<AssertPlaceHolder>" ; } getTableCellRendererComponent ( kg . apc . jmeter . vizualizers . JTable , java . lang . Object , boolean , boolean , int , int ) { if ( table != null ) { javax . swing . table . JTableHeader header = table . getTableHeader ( ) ; if ( header != null ) { setForeground ( header . getForeground ( ) ) ; setBackground ( header . getBackground ( ) ) ; setFont ( header . getFont ( ) ) ; } setText ( getText ( value , row , column ) ) ; setBorder ( kg . apc . jmeter . vizualizers . UIManager . getBorder ( "TableHeader.cellBorder" ) ) ; setHorizontalAlignment ( SwingConstants . CENTER ) ; } return this ; }
org . junit . Assert . assertTrue ( ( result instanceof kg . apc . jmeter . graphs . HeaderAsTextRenderer ) )
populatedStreamMetadataBuilderShouldBeMutable ( ) { com . github . msemys . esjc . StreamMetadata streamMetadata = com . github . msemys . esjc . StreamMetadata . newBuilder ( ) . maxCount ( 19L ) . maxAge ( java . time . Duration . ofSeconds ( 82 ) ) . truncateBefore ( 8L ) . cacheControl ( java . time . Duration . ofSeconds ( 17 ) ) . aclReadRoles ( asList ( "customBoolean" 1 , "kyle" , "b" 5 , "b" 1 ) ) . aclWriteRoles ( asList ( "b" 3 ) ) . aclDeleteRoles ( asList ( "customBoolean" 2 ) ) . aclMetaReadRoles ( asList ( "b" 6 , "mackey" ) ) . aclMetaWriteRoles ( asList ( "customBoolean" 4 ) ) . customProperty ( "b" 9 , "b" 7 ) . customProperty ( "customInt" , ( - 179 ) ) . customProperty ( "customDouble" , 1.7 ) . customProperty ( "customLong" , 123123123123123123L ) . customProperty ( "customBoolean" , true ) . customProperty ( "customStringArray" , "b" 0 , "b" , "b" 8 , null ) . customProperty ( "customBoolean" 3 , 1 , 2 , 3 , null ) . customProperty ( "customBoolean" 0 , 111111111111111111L , 222222222222222222L , 333333333333333333L , null ) . build ( ) ; com . github . msemys . esjc . StreamMetadata expectedStreamMetadata = com . github . msemys . esjc . StreamMetadata . newBuilder ( ) . maxCount ( 19L ) . maxAge ( java . time . Duration . ofSeconds ( 17 ) ) . truncateBefore ( 8L ) . cacheControl ( java . time . Duration . ofDays ( 82 ) ) . aclReadRoles ( asList ( "foo" ) ) . aclWriteRoles ( asList ( "bar" ) ) . aclDeleteRoles ( asList ( "b" 4 ) ) . customProperty ( "b" 9 , "b" 7 ) . customProperty ( "customInt" , 123 ) . customProperty ( "customDouble" , 2.8 ) . customProperty ( "customLong" , 4444444444444444444L ) . customProperty ( "customBoolean" , false ) . customProperty ( "customStringArray" , "b" 0 , "b" , "b" 8 , "b" 2 ) . customProperty ( "customBoolean" 3 , 1 , 2 , 3 , 4 ) . customProperty ( "customBoolean" 0 , ( ( java . lang . Long [ ] ) ( null ) ) ) . build ( ) ; com . github . msemys . esjc . StreamMetadata result = streamMetadata . toBuilder ( ) . maxAge ( java . time . Duration . ofSeconds ( 17 ) ) . cacheControl ( java . time . Duration . ofDays ( 82 ) ) . aclReadRoles ( asList ( "foo" ) ) . aclWriteRoles ( asList ( "bar" ) ) . aclDeleteRoles ( asList ( "b" 4 ) ) . aclMetaReadRoles ( null ) . aclMetaWriteRoles ( null ) . customProperty ( "customInt" , 123 ) . customProperty ( "customDouble" , 2.8 ) . customProperty ( "customLong" , 4444444444444444444L ) . customProperty ( "customBoolean" , false ) . customProperty ( "customStringArray" , "b" 0 , "b" , "b" 8 , "b" 2 ) . customProperty ( "customBoolean" 3 , 1 , 2 , 3 , 4 ) . customProperty ( "customBoolean" 0 , ( ( java . lang . Long [ ] ) ( null ) ) ) . build ( ) ; "<AssertPlaceHolder>" ; } toJson ( ) { return com . github . msemys . esjc . SystemSettings . gson . toJson ( this ) ; }
org . junit . Assert . assertEquals ( expectedStreamMetadata . toJson ( ) , result . toJson ( ) )
testHornerForm1 ( ) { org . apache . commons . math3 . random . RandomGenerator rnd = getRandom ( ) ; org . apache . commons . math3 . random . RandomDataGenerator rndd = getRandomData ( ) ; cc . redberry . rings . IntegersZp64 ring = cc . redberry . rings . Rings . Zp64 ( cc . redberry . rings . primes . SmallPrimes . nextPrime ( ( 1 << 15 ) ) ) ; org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics hornerStat = new org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics ( ) ; org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics hornerCreateStat = new org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics ( ) ; org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics plainStat = new org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics ( ) ; long start ; long elapsed ; int nIterations = 100 ; int nEvaluations = its ( 10 , 10 ) ; int nVars = 5 ; int minDeg = 3250 ; int minSize = 1000 ; int [ ] varsSeq = cc . redberry . rings . util . ArraysUtil . sequence ( 0 , nVars ) ; for ( int i = 0 ; i < nIterations ; ++ i ) { if ( i == ( nIterations / 10 ) ) java . util . Arrays . asList ( hornerStat , hornerCreateStat , plainStat ) . forEach ( DescriptiveStatistics :: clear ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . MultivariatePolynomialZp64 p = cc . redberry . rings . poly . multivar . RandomMultivariatePolynomials . randomPolynomial ( nVars , rndd . nextInt ( minDeg , ( 2 * minDeg ) ) , rndd . nextInt ( minSize , ( 2 * minSize ) ) , ring , rnd ) ; int [ ] variables = new int [ 3 + ( rnd . nextInt ( ( ( p . nVariables ) - 3 ) ) ) ] ; long [ ] values = new long [ variables . length ] ; cc . redberry . rings . util . ArraysUtil . shuffle ( varsSeq , rnd ) ; java . lang . System . arraycopy ( varsSeq , 0 , variables , 0 , values . length ) ; start = java . lang . System . nanoTime ( ) ; cc . redberry . rings . poly . multivar . HornerFormZp64 hornerForm = p . getHornerForm ( variables ) ; elapsed = ( java . lang . System . nanoTime ( ) ) - start ; hornerStat . addValue ( elapsed ) ; hornerCreateStat . addValue ( elapsed ) ; for ( int nEval = 0 ; nEval < nEvaluations ; ++ nEval ) { for ( int j = 0 ; j < ( values . length ) ; j ++ ) values [ j ] = ring . randomElement ( rnd ) ; start = java . lang . System . nanoTime ( ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . MultivariatePolynomialZp64 horner = hornerForm . evaluate ( values ) ; elapsed = ( java . lang . System . nanoTime ( ) ) - start ; hornerStat . addValue ( elapsed ) ; start = java . lang . System . nanoTime ( ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . MultivariatePolynomialZp64 plain = p . eliminate ( variables , values ) ; elapsed = ( java . lang . System . nanoTime ( ) ) - start ; plainStat . addValue ( elapsed ) ; "<AssertPlaceHolder>" ; } } System . out . println ( ( "Horner<sp>:<sp>" + ( statisticsNanotime ( hornerStat ) ) ) ) ; System . out . println ( ( "Horner<sp>create<sp>:<sp>" + ( statisticsNanotime ( hornerCreateStat ) ) ) ) ; System . out . println ( ( "Plain<sp>:<sp>" + ( statisticsNanotime ( plainStat ) ) ) ) ; } eliminate ( int , long ) { value = ring . modulus ( value ) ; cc . redberry . rings . poly . multivar . MonomialSet < cc . redberry . rings . poly . multivar . MonomialZp64 > newData = new cc . redberry . rings . poly . multivar . MonomialSet ( ordering ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . lPrecomputedPowers powers = new cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . lPrecomputedPowers ( value , ring ) ; for ( cc . redberry . rings . poly . multivar . MonomialZp64 el : terms ) { long val = ring . multiply ( el . coefficient , powers . pow ( el . exponents [ variable ] ) ) ; add ( newData , el . without ( variable ) . setCoefficient ( val ) ) ; } return new cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 ( ( ( nVariables ) - 1 ) , ring , ordering , newData ) ; }
org . junit . Assert . assertEquals ( plain , horner )
getAttributePrefix ( ) { java . lang . String expected = "expected" ; int index = 10 ; when ( this . reader . getAttributePrefix ( index ) ) . thenReturn ( expected ) ; java . lang . String result = this . filter . getAttributePrefix ( index ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertSame ( expected , result )
entryByLicenseId ( ) { java . io . File repoFile = null ; try { sernet . verinice . model . licensemanagement . LicenseManagementEntry entry = getSingleCryptedEntry ( ) ; repoFile = addLicenseToRepository ( entry ) ; sernet . verinice . model . licensemanagement . LicenseManagementEntry entryFromRepo = licenseManagementService . getLicenseEntryForLicenseId ( entry . getLicenseID ( ) , false ) ; "<AssertPlaceHolder>" ; } finally { org . apache . commons . io . FileUtils . forceDelete ( repoFile ) ; } } equals ( java . lang . Object ) { if ( ! ( entry instanceof sernet . verinice . report . service . impl . TocHelper2 . TocEntry ) ) { return false ; } return this . toString ( ) . equals ( ( ( sernet . verinice . report . service . impl . TocHelper2 . TocEntry < TITLE , PAGENUMBER > ) ( entry ) ) . toString ( ) ) ; }
org . junit . Assert . assertTrue ( entry . equals ( entryFromRepo ) )
testExceptionThrow ( ) { context . checking ( new org . jmock . Expectations ( ) { { oneOf ( securityContext ) . getAuthentication ( ) ; will ( throwException ( new java . lang . Exception ( ) ) ) ; } } ) ; "<AssertPlaceHolder>" ; } runTest ( boolean ) { org . springframework . security . context . SecurityContext originalSecurityContext = org . springframework . security . context . SecurityContextHolder . getContext ( ) ; org . springframework . security . context . SecurityContextHolder . setContext ( securityContext ) ; org . eurekastreams . server . action . principal . SpringSecurityContextPrincipalPopulator sut = new org . eurekastreams . server . action . principal . SpringSecurityContextPrincipalPopulator ( exceptionOnError ) ; org . eurekastreams . commons . actions . context . Principal result ; try { result = sut . transform ( request ) ; } finally { org . springframework . security . context . SecurityContextHolder . setContext ( originalSecurityContext ) ; } context . assertIsSatisfied ( ) ; return result ; }
org . junit . Assert . assertNull ( runTest ( true ) )
uniquify ( ) { int [ ] ints = new int [ ] { 1 , 4 , 4 , 2 , 7 , 3 , 8 , 0 , 0 , 3 } ; int [ ] actual = biweekly . util . com . google . ical . iter . Util . uniquify ( ints ) ; int [ ] expected = new int [ ] { 0 , 1 , 2 , 3 , 4 , 7 , 8 } ; "<AssertPlaceHolder>" ; } uniquify ( int [ ] ) { biweekly . util . com . google . ical . iter . IntSet iset = new biweekly . util . com . google . ical . iter . IntSet ( ) ; for ( int i : ints ) { iset . add ( i ) ; } return iset . toIntArray ( ) ; }
org . junit . Assert . assertTrue ( java . util . Arrays . equals ( expected , actual ) )
testNotifiesAboutEachFileVersionDeletion ( ) { _withMaximumNumberOfVersionsConfigured ( 2 , ( ) -> { com . liferay . portal . kernel . service . ServiceContext serviceContext = com . liferay . portal . kernel . test . util . ServiceContextTestUtil . getServiceContext ( _group . getGroupId ( ) ) ; com . liferay . portal . kernel . repository . model . FileEntry fileEntry = _addRandomFileEntry ( serviceContext ) ; List < com . liferay . portal . kernel . repository . model . FileVersion > deletedFileVersions = new com . liferay . document . library . repository . capabilities . test . ArrayList < > ( ) ; com . liferay . registry . Registry registry = com . liferay . registry . RegistryUtil . getRegistry ( ) ; ServiceRegistration < com . liferay . document . library . versioning . VersionPurger . VersionPurgedListener > capabilityServiceRegistration = registry . registerService ( . class , deletedFileVersions :: add ) ; try { for ( int i = 0 ; i < 10 ; i ++ ) { _generateNewVersion ( fileEntry , serviceContext ) ; } "<AssertPlaceHolder>" ; } finally { capabilityServiceRegistration . unregister ( ) ; } } ) ; } toString ( ) { com . liferay . petra . string . StringBundler sb = new com . liferay . petra . string . StringBundler ( 23 ) ; sb . append ( ",<sp>width=" 1 ) ; sb . append ( uuid ) ; sb . append ( ",<sp>width=" 0 ) ; sb . append ( amImageEntryId ) ; sb . append ( ",<sp>groupId=" ) ; sb . append ( groupId ) ; sb . append ( ",<sp>companyId=" ) ; sb . append ( companyId ) ; sb . append ( ",<sp>createDate=" ) ; sb . append ( createDate ) ; sb . append ( ",<sp>configurationUuid=" ) ; sb . append ( configurationUuid ) ; sb . append ( ",<sp>fileVersionId=" ) ; sb . append ( fileVersionId ) ; sb . append ( ",<sp>mimeType=" ) ; sb . append ( mimeType ) ; sb . append ( ",<sp>height=" ) ; sb . append ( height ) ; sb . append ( ",<sp>width=" ) ; sb . append ( width ) ; sb . append ( ",<sp>size=" ) ; sb . append ( size ) ; sb . append ( "}" ) ; return sb . toString ( ) ; }
org . junit . Assert . assertEquals ( deletedFileVersions . toString ( ) , 9 , deletedFileVersions . size ( ) )
getIotHubStatusCodeMapsHubOrDeviceNotFoundCorrectly ( ) { final int httpsStatus = 404 ; com . microsoft . azure . sdk . iot . device . IotHubStatusCode testStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . getIotHubStatusCode ( httpsStatus ) ; final com . microsoft . azure . sdk . iot . device . IotHubStatusCode expectedStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . HUB_OR_DEVICE_ID_NOT_FOUND ; "<AssertPlaceHolder>" ; } getIotHubStatusCode ( int ) { com . microsoft . azure . sdk . iot . device . IotHubStatusCode iotHubStatus ; switch ( httpsStatus ) { case 200 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . OK ; break ; case 204 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . OK_EMPTY ; break ; case 400 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . BAD_FORMAT ; break ; case 401 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . UNAUTHORIZED ; break ; case 403 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . TOO_MANY_DEVICES ; break ; case 404 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . HUB_OR_DEVICE_ID_NOT_FOUND ; break ; case 412 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . PRECONDITION_FAILED ; break ; case 413 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . REQUEST_ENTITY_TOO_LARGE ; break ; case 429 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . THROTTLED ; break ; case 500 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . INTERNAL_SERVER_ERROR ; break ; case 503 : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . SERVER_BUSY ; break ; default : iotHubStatus = com . microsoft . azure . sdk . iot . device . IotHubStatusCode . ERROR ; } return iotHubStatus ; }
org . junit . Assert . assertThat ( testStatus , org . hamcrest . CoreMatchers . is ( expectedStatus ) )
testTaobao ( ) { io . github . bonigarcia . wdm . WebDriverManager . chromedriver ( ) . config ( ) . setAvoidAutoVersion ( true ) . setChromeDriverMirrorUrl ( new java . net . URL ( "http://npm.taobao.org/mirrors/chromedriver/" ) ) ; io . github . bonigarcia . wdm . WebDriverManager . chromedriver ( ) . useMirror ( ) . forceDownload ( ) . setup ( ) ; java . io . File binary = new java . io . File ( io . github . bonigarcia . wdm . WebDriverManager . chromedriver ( ) . getBinaryPath ( ) ) ; "<AssertPlaceHolder>" ; } getBinaryPath ( ) { return resolve ( binaryPath ) ; }
org . junit . Assert . assertTrue ( binary . exists ( ) )
testSourceEntity ( ) { org . apache . cayenne . map . Relationship rel = new org . apache . cayenne . map . MockRelationship ( ) ; org . apache . cayenne . map . Entity tstEntity = new org . apache . cayenne . map . MockEntity ( ) ; rel . setSourceEntity ( tstEntity ) ; "<AssertPlaceHolder>" ; } getSourceEntity ( ) { return sourceEntity ; }
org . junit . Assert . assertSame ( tstEntity , rel . getSourceEntity ( ) )
testDynamicQueryByPrimaryKeyMissing ( ) { com . liferay . portal . kernel . dao . orm . DynamicQuery dynamicQuery = com . liferay . portal . kernel . dao . orm . DynamicQueryFactoryUtil . forClass ( com . liferay . dynamic . data . mapping . model . DDMFormInstanceVersion . class , _dynamicQueryClassLoader ) ; dynamicQuery . add ( com . liferay . portal . kernel . dao . orm . RestrictionsFactoryUtil . eq ( "formInstanceVersionId" , com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ) ) ; java . util . List < com . liferay . dynamic . data . mapping . model . DDMFormInstanceVersion > result = _persistence . findWithDynamicQuery ( dynamicQuery ) ; "<AssertPlaceHolder>" ; } size ( ) { if ( ( _workflowTaskAssignees ) != null ) { return _workflowTaskAssignees . size ( ) ; } return _kaleoTaskAssignmentInstanceLocalService . getKaleoTaskAssignmentInstancesCount ( _kaleoTaskInstanceToken . getKaleoTaskInstanceTokenId ( ) ) ; }
org . junit . Assert . assertEquals ( 0 , result . size ( ) )
removeWorks ( ) { for ( java . util . Map < java . lang . Integer , java . lang . Integer > any : net . java . quickcheck . generator . CombinedGeneratorsIterables . someMaps ( net . java . quickcheck . generator . PrimitiveGenerators . integers ( ) , net . java . quickcheck . generator . PrimitiveGenerators . integers ( ) ) ) { com . google . firebase . database . collection . ImmutableSortedMap < java . lang . Integer , java . lang . Integer > map = com . google . firebase . database . collection . ArraySortedMap . fromMap ( any , com . google . firebase . database . collection . ArraySortedMapTest . IntComparator ) ; for ( Map . Entry < java . lang . Integer , java . lang . Integer > entry : any . entrySet ( ) ) { map = map . remove ( entry . getKey ( ) ) ; } "<AssertPlaceHolder>" ; } } size ( ) { return ( this . end ) - ( this . start ) ; }
org . junit . Assert . assertEquals ( 0 , map . size ( ) )
shouldBeIterable ( ) { com . qcadoo . commons . functional . LazyStream < java . lang . Integer > stream = com . qcadoo . commons . functional . LazyStream . create ( 1 , com . qcadoo . commons . functional . LazyStreamTest . INCREMENT_BY_ONE ) ; java . util . List < java . lang . Integer > iterableValues = com . google . common . collect . FluentIterable . from ( stream ) . limit ( 5 ) . toList ( ) ; "<AssertPlaceHolder>" ; } create ( U , com . google . common . base . Function ) { return new com . qcadoo . commons . functional . LazyStream < U > ( firstElement , inductionStep ) ; }
org . junit . Assert . assertEquals ( com . google . common . collect . Lists . newArrayList ( 1 , 2 , 3 , 4 , 5 ) , iterableValues )
testGetMissingRomFilesForSoftwareClone ( ) { org . tibennetwork . iarcade . mame . FakeMameRuntime mame = new org . tibennetwork . iarcade . mame . FakeMameRuntime ( ) ; java . util . List < java . io . InputStream > inputStreams = new java . util . ArrayList ( ) ; inputStreams . add ( new java . io . FileInputStream ( "src/test/resources/xml/snes.xml" ) ) ; inputStreams . add ( new java . io . FileInputStream ( "src/test/resources/xml/snes-sl.xml" ) ) ; inputStreams . add ( new java . io . FileInputStream ( "src/test/resources/xml/snes-sl.xml" ) ) ; mame . setInputStreamsToReturn ( inputStreams ) ; org . tibennetwork . iarcade . mame . MachineRepository mr = new org . tibennetwork . iarcade . mame . MachineRepository ( mame ) ; org . tibennetwork . iarcade . mame . Machine m = mr . findByName ( "snes" ) ; org . tibennetwork . iarcade . mame . SoftwareRepository sr = new org . tibennetwork . iarcade . mame . SoftwareRepository ( mame ) ; org . tibennetwork . iarcade . mame . Software s = sr . findByMachineAndByName ( m , "megamnx2u" ) ; java . util . Set < java . io . File > romPaths = new java . util . HashSet ( ) ; romPaths . add ( new java . io . File ( "src/test/resources/full-rompath" ) ) ; java . util . Set < org . tibennetwork . iarcade . mame . SoftwareRomFile > expectedMissingRomFiles = new java . util . HashSet ( ) ; expectedMissingRomFiles . add ( new org . tibennetwork . iarcade . mame . SoftwareRomFile ( "snes" , "megamnx2u" ) ) ; expectedMissingRomFiles . add ( new org . tibennetwork . iarcade . mame . SoftwareRomFile ( "snes" , "megamnx2" ) ) ; "<AssertPlaceHolder>" ; } getMissingRomFiles ( java . util . Set ) { java . util . Set < org . tibennetwork . iarcade . mame . SoftwareRomFile > missingRomFiles = new java . util . HashSet ( ) ; softwareFilesLoop : for ( org . tibennetwork . iarcade . mame . SoftwareRomFile file : this . getNeededRomFiles ( ) ) { for ( java . io . File romPath : romPaths ) { java . io . File zippedFileInRomPath = new java . io . File ( ( ( ( romPath . getAbsolutePath ( ) ) + ( java . io . File . separator ) ) + ( file . getZipRelativePath ( ) ) ) ) ; java . io . File sevenZippedFileInRomPath = new java . io . File ( ( ( ( romPath . getAbsolutePath ( ) ) + ( java . io . File . separator ) ) + ( file . getSevenZipRelativePath ( ) ) ) ) ; if ( ( zippedFileInRomPath . exists ( ) ) || ( sevenZippedFileInRomPath . exists ( ) ) ) { continue softwareFilesLoop ; } } missingRomFiles . add ( file ) ; } return missingRomFiles ; }
org . junit . Assert . assertThat ( s . getMissingRomFiles ( romPaths ) , org . hamcrest . CoreMatchers . equalTo ( expectedMissingRomFiles ) )
testDoGet ( ) { uk . gov . dstl . baleen . testing . servlets . ServletCaller caller = new uk . gov . dstl . baleen . testing . servlets . ServletCaller ( ) ; javax . servlet . http . HttpServletResponse response = caller . doGet ( servlet ) ; "<AssertPlaceHolder>" ; verify ( servlet ) . service ( caller . getRequest ( ) , caller . getResponse ( ) ) ; } doGet ( javax . servlet . http . HttpServlet ) { return doMethod ( "GET" , servlet ) ; }
org . junit . Assert . assertNotNull ( response )
testRetainOverflow ( ) { io . netty . buffer . AbstractReferenceCountedByteBuf referenceCounted = io . netty . buffer . AbstractReferenceCountedByteBufTest . newReferenceCounted ( ) ; referenceCounted . setRefCnt ( Integer . MAX_VALUE ) ; "<AssertPlaceHolder>" ; referenceCounted . retain ( ) ; } refCnt ( ) { return buf . refCnt ( ) ; }
org . junit . Assert . assertEquals ( Integer . MAX_VALUE , referenceCounted . refCnt ( ) )
data_should_be_compared_between_expected_and_current_data ( ) { com . lordofthejars . nosqlunit . mongodb . MongoOperation mongoOperation = new com . lordofthejars . nosqlunit . mongodb . MongoOperation ( com . lordofthejars . nosqlunit . mongodb . InMemoryMongoDbConfigurationBuilder . inMemoryMongoDb ( ) . databaseName ( "test" ) . build ( ) ) ; mongoOperation . insert ( new java . io . ByteArrayInputStream ( com . lordofthejars . nosqlunit . mongodb . WhenEmbeddedMongoDbOperationsArRequired . DATA . getBytes ( ) ) ) ; boolean result = mongoOperation . databaseIs ( new java . io . ByteArrayInputStream ( com . lordofthejars . nosqlunit . mongodb . WhenEmbeddedMongoDbOperationsArRequired . DATA . getBytes ( ) ) ) ; "<AssertPlaceHolder>" ; } databaseIs ( java . io . InputStream ) { return compareData ( contentStream ) ; }
org . junit . Assert . assertThat ( result , org . hamcrest . CoreMatchers . is ( true ) )
testGetReferencedComponentId ( ) { java . lang . String expectedStringValue = "SomeStringValue" ; java . lang . String referencedComponentId ; closeProperties . referencedComponent . componentInstanceId . setValue ( expectedStringValue ) ; referencedComponentId = closeProperties . getReferencedComponentId ( ) ; "<AssertPlaceHolder>" ; } getReferencedComponentId ( ) { return referencedComponentId ; }
org . junit . Assert . assertEquals ( referencedComponentId , expectedStringValue )
testGetInstance ( ) { org . junit . Assume . assumeTrue ( ( ( org . apache . hadoop . crypto . OpensslCipher . getLoadingFailureReason ( ) ) == null ) ) ; org . apache . hadoop . crypto . OpensslCipher cipher = org . apache . hadoop . crypto . OpensslCipher . getInstance ( "AES/CTR/NoPadding" ) ; "<AssertPlaceHolder>" ; try { cipher = org . apache . hadoop . crypto . OpensslCipher . getInstance ( "AES2/CTR/NoPadding" ) ; org . junit . Assert . fail ( "Should<sp>specify<sp>correct<sp>algorithm." ) ; } catch ( java . security . NoSuchAlgorithmException e ) { } try { cipher = org . apache . hadoop . crypto . OpensslCipher . getInstance ( "AES/CTR/NoPadding2" ) ; org . junit . Assert . fail ( "Should<sp>specify<sp>correct<sp>padding." ) ; } catch ( javax . crypto . NoSuchPaddingException e ) { } } getInstance ( java . lang . String ) { org . apache . hadoop . crypto . OpensslCipher . Transform transform = org . apache . hadoop . crypto . OpensslCipher . tokenizeTransformation ( transformation ) ; int algMode = org . apache . hadoop . crypto . OpensslCipher . AlgMode . get ( transform . alg , transform . mode ) ; int padding = org . apache . hadoop . crypto . OpensslCipher . Padding . get ( transform . padding ) ; long context = org . apache . hadoop . crypto . OpensslCipher . initContext ( algMode , padding ) ; return new org . apache . hadoop . crypto . OpensslCipher ( context , algMode , padding ) ; }
org . junit . Assert . assertTrue ( ( cipher != null ) )
readData_InvalidLocationEvent ( ) { org . mockito . Mockito . when ( filereader . fileExist ( org . mockito . Mockito . anyString ( ) ) ) . thenReturn ( true ) ; org . mockito . Mockito . when ( filereader . readAllLine ( org . mockito . Mockito . anyString ( ) ) ) . thenReturn ( new java . lang . String [ ] { "1.487031029351E9<sp>" , "1.487031045307E9<sp>47.7943242<sp>-122.2030202<sp>Loction<sp>network" } ) ; java . util . List < com . att . aro . core . peripheral . pojo . LocationEvent > listLocationEvent = traceDataReader . readData ( traceFolder , 0.0 ) ; "<AssertPlaceHolder>" ; } size ( ) { return sessionTable . size ( ) ; }
org . junit . Assert . assertEquals ( 1.0 , listLocationEvent . size ( ) , 0 )
shouldReturnFalseIfFileIsNotJarFile ( ) { com . openpojo . reflection . PojoClass pojoClass = com . openpojo . reflection . impl . PojoClassFactory . getPojoClass ( this . getClass ( ) ) ; java . lang . String sourcePath = new java . net . URL ( pojoClass . getSourcePath ( ) ) . getPath ( ) ; com . openpojo . reflection . java . packageloader . reader . JarFileReader jarFileReader = com . openpojo . reflection . java . packageloader . reader . JarFileReader . getInstance ( sourcePath ) ; "<AssertPlaceHolder>" ; } isValid ( ) { return ( ( getPackageLoaders ( ) . size ( ) ) > 0 ) || ( com . openpojo . reflection . java . packageloader . Package . JAVA_CLASSPATH_CLASS_LOADER . hasPackage ( packageName ) ) ; }
org . junit . Assert . assertFalse ( jarFileReader . isValid ( ) )
completedAndSelectedByDateCriteriaShouldBeDeleted ( ) { newJobStatus ( ) . startTime ( new java . util . Date ( ) ) . finishTime ( org . candlepin . util . Util . tomorrow ( ) ) . create ( ) ; this . curator . cleanUpOldCompletedJobs ( org . candlepin . util . Util . addDaysToDt ( 2 ) ) ; "<AssertPlaceHolder>" ; } listAll ( ) { org . hibernate . criterion . DetachedCriteria criteria = this . createSecureDetachedCriteria ( ) ; return this . cpQueryFactory . < E > buildQuery ( this . currentSession ( ) , criteria ) ; }
org . junit . Assert . assertEquals ( 0 , this . curator . listAll ( ) . list ( ) . size ( ) )
testCreateManagerUrlForDefaultHostAndPort ( ) { final java . net . URL actualManagerUrl = commonTomcatConfiguration . createManagerUrl ( ) ; final java . lang . String actualManagerUrlString = actualManagerUrl . toString ( ) ; final java . lang . String expectedManagerUrlString = "http://localhost:8080/manager" ; "<AssertPlaceHolder>" ; } createManagerUrl ( ) { try { final java . lang . String template = "http://%s:%d/manager" ; final java . lang . String urlString = java . lang . String . format ( template , bindAddress , bindHttpPort ) ; return new java . net . URL ( urlString ) ; } catch ( final java . net . MalformedURLException e ) { throw new org . jboss . arquillian . container . spi . ConfigurationException ( "Manager<sp>URL<sp>is<sp>not<sp>valid,<sp>please<sp>provide<sp>" , e ) ; } }
org . junit . Assert . assertEquals ( expectedManagerUrlString , actualManagerUrlString )
testTriggerScanStarting ( ) { com . sonyericsson . hudson . plugins . gerrit . trigger . hudsontrigger . actions . manual . TriggerMonitor monitor = new com . sonyericsson . hudson . plugins . gerrit . trigger . hudsontrigger . actions . manual . TriggerMonitor ( ) ; com . sonyericsson . hudson . plugins . gerrit . trigger . events . ManualPatchsetCreated patch = com . sonyericsson . hudson . plugins . gerrit . trigger . mock . Setup . createManualPatchsetCreated ( ) ; monitor . add ( patch ) ; monitor . triggerScanStarting ( patch ) ; com . sonyericsson . hudson . plugins . gerrit . trigger . hudsontrigger . actions . manual . TriggerMonitor . EventState state = monitor . getEvents ( ) . get ( 0 ) ; "<AssertPlaceHolder>" ; } isTriggerScanStarted ( ) { return triggerScanStarted ; }
org . junit . Assert . assertTrue ( state . isTriggerScanStarted ( ) )
testAsyncCreateInstance_noControllerSet ( ) { final org . oscm . provisioning . data . InstanceRequest rq = new org . oscm . provisioning . data . InstanceRequest ( ) ; rq . setOrganizationId ( "org123" ) ; rq . setOrganizationName ( "Fujitsu" ) ; rq . setSubscriptionId ( "sub123" ) ; rq . setDefaultLocale ( "de" ) ; rq . setLoginUrl ( "http://bes/" ) ; final org . oscm . provisioning . data . ServiceParameter param = new org . oscm . provisioning . data . ServiceParameter ( ) ; param . setParameterId ( "appParam1" ) ; param . setValue ( "appValue1" ) ; rq . setParameterValue ( new java . util . ArrayList < org . oscm . provisioning . data . ServiceParameter > ( ) ) ; rq . getParameterValue ( ) . add ( param ) ; org . oscm . provisioning . data . BaseResult result = proxy . asyncCreateInstance ( rq , null ) ; "<AssertPlaceHolder>" ; } getRc ( ) { return localRc ; }
org . junit . Assert . assertFalse ( ( ( result . getRc ( ) ) == 0 ) )
testGetAttributeValueList ( ) { org . finra . herd . model . api . xml . AttributeValueListKey attributeValueListKey = new org . finra . herd . model . api . xml . AttributeValueListKey ( ATTRIBUTE_VALUE_LIST_NAMESPACE , ATTRIBUTE_VALUE_LIST_NAME ) ; org . finra . herd . model . api . xml . AttributeValueList attributeValueList = new org . finra . herd . model . api . xml . AttributeValueList ( ATTRIBUTE_VALUE_LIST_ID , attributeValueListKey ) ; when ( attributeValueListService . getAttributeValueList ( attributeValueListKey ) ) . thenReturn ( attributeValueList ) ; org . finra . herd . model . api . xml . AttributeValueList result = attributeValueListRestController . getAttributeValueList ( org . finra . herd . rest . ATTRIBUTE_VALUE_LIST_NAMESPACE , org . finra . herd . rest . ATTRIBUTE_VALUE_LIST_NAME ) ; verify ( attributeValueListService ) . getAttributeValueList ( attributeValueListKey ) ; verifyNoMoreInteractions ( attributeValueListService ) ; "<AssertPlaceHolder>" ; } getAttributeValueList ( org . finra . herd . model . api . xml . AttributeValueListKey ) { attributeValueListHelper . validateAttributeValueListKey ( attributeValueListKey ) ; org . finra . herd . model . jpa . AttributeValueListEntity attributeValueListEntity = attributeValueListDaoHelper . getAttributeValueListEntity ( attributeValueListKey ) ; return attributeValueListDaoHelper . createAttributeValueListFromEntity ( attributeValueListEntity ) ; }
org . junit . Assert . assertEquals ( attributeValueList , result )
testSetsForegroundWithRGB ( ) { com . eclipsesource . tabris . ui . UIConfiguration configuration = new com . eclipsesource . tabris . ui . UIConfiguration ( ) ; configuration . setForeground ( 233 , 233 , 233 ) ; org . eclipse . swt . graphics . RGB actualForeground = configuration . getForeground ( ) ; "<AssertPlaceHolder>" ; } getForeground ( ) { return foreground ; }
org . junit . Assert . assertEquals ( new org . eclipse . swt . graphics . RGB ( 233 , 233 , 233 ) , actualForeground )
parallelStream_whenWorks_thenCorrect ( ) { com . baeldung . java8 . Stream < com . baeldung . stream . Product > streamOfCollection = productList . parallelStream ( ) ; boolean isParallel = streamOfCollection . isParallel ( ) ; boolean haveBigPrice = streamOfCollection . map ( ( product ) -> ( product . getPrice ( ) ) * 12 ) . anyMatch ( ( price ) -> price > 200 ) ; "<AssertPlaceHolder>" ; } getPrice ( ) { return price ; }
org . junit . Assert . assertTrue ( ( isParallel && haveBigPrice ) )
openResource ( ) { org . apache . lucene . analysis . util . ClasspathResourceLoader parent = new org . apache . lucene . analysis . util . ClasspathResourceLoader ( getClass ( ) . getClassLoader ( ) ) ; com . s24 . search . solr . analysis . jdbc . JdbcReader reader = new com . s24 . search . solr . analysis . jdbc . JdbcResourceLoaderTest . TestJdbcReader ( "test=>test1,test2" ) ; java . nio . charset . Charset charset = java . nio . charset . Charset . forName ( "UTF-8" ) ; com . s24 . search . solr . analysis . jdbc . JdbcResourceLoader loader = new com . s24 . search . solr . analysis . jdbc . JdbcResourceLoader ( parent , reader , charset ) ; java . io . InputStream resource = loader . openResource ( JdbcResourceLoader . DATABASE ) ; java . io . StringWriter writer = new java . io . StringWriter ( ) ; org . apache . commons . io . IOUtils . copy ( resource , writer , charset ) ; "<AssertPlaceHolder>" ; } openResource ( java . lang . String ) { if ( com . s24 . search . solr . analysis . jdbc . JdbcResourceLoader . DATABASE . equals ( resource ) ) { return new org . apache . commons . io . input . ReaderInputStream ( reader . getReader ( ) , charset ) ; } return parent . openResource ( resource ) ; }
org . junit . Assert . assertEquals ( "test=>test1,test2" , writer . toString ( ) )
test_getAllAddressTags ( ) { "<AssertPlaceHolder>" ; } getAllAddressTags ( ) { return io . github . data4all . model . data . Tags . ADDRESS_TAG_LIST ; }
org . junit . Assert . assertEquals ( 5 , io . github . data4all . model . data . Tags . getAllAddressTags ( ) . size ( ) )
testAll ( ) { append ( "[groups]" , globalPolicyFile ) ; append ( "group<sp>=<sp>malicious_role" , globalPolicyFile ) ; append ( "[roles]" , globalPolicyFile ) ; append ( "malicious_role<sp>=<sp>*" , globalPolicyFile ) ; org . apache . sentry . policy . common . PolicyEngine policy = org . apache . sentry . policy . hive . DBPolicyTestUtil . createPolicyEngineForTest ( "server1" , globalPolicyFile . getPath ( ) ) ; com . google . common . collect . ImmutableSet < java . lang . String > permissions = policy . getAllPrivileges ( com . google . common . collect . Sets . newHashSet ( "group" ) , ActiveRoleSet . ALL ) ; "<AssertPlaceHolder>" ; } toString ( ) { return SentryConstants . AUTHORIZABLE_JOINER . join ( parts ) ; }
org . junit . Assert . assertTrue ( permissions . toString ( ) , permissions . isEmpty ( ) )
testConfigure ( ) { System . out . println ( "Testing<sp>RoundRobinChannelSelector.configure" ) ; channelSelector . configure ( context ) ; "<AssertPlaceHolder>" ; } configure ( org . apache . flume . Context ) { consumerKey = context . getString ( "maxBatchSize" 6 ) ; consumerSecret = context . getString ( "consumerSecret" ) ; accessToken = context . getString ( "consumerSecret" 1 ) ; accessTokenSecret = context . getString ( "maxBatchSize" 0 ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( "Consumer<sp>Key:<sp>'" + ( consumerKey ) ) + "maxBatchSize" 4 ) ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( "maxBatchSize" 3 + ( consumerSecret ) ) + "maxBatchSize" 4 ) ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( "maxBatchSize" 5 + ( accessToken ) ) + "maxBatchSize" 4 ) ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( "Access<sp>Token<sp>Secret:<sp>'" + ( accessTokenSecret ) ) + "maxBatchSize" 4 ) ) ; java . lang . String southWestLatitude ; java . lang . String southWestLongitude ; java . lang . String northEastLatitude ; java . lang . String northEastLongitude ; java . lang . String keywords ; southWestLatitude = context . getString ( "south_west_latitude" ) ; southWestLongitude = context . getString ( "maxBatchSize" 1 ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( ( ( "maxBatchSize" 8 + southWestLatitude ) + "maxBatchSize" 2 ) + southWestLongitude ) + "maxBatchSize" 4 ) ) ; northEastLatitude = context . getString ( "north_east_latitude" ) ; northEastLongitude = context . getString ( "north_east_longitude" ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( ( ( "North-East<sp>coordinate:<sp>'" + northEastLatitude ) + "maxBatchSize" 2 ) + northEastLongitude ) + "maxBatchSize" 4 ) ) ; keywords = context . getString ( "keywords" ) ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( "maxBatchSize" 7 + keywords ) + "maxBatchSize" 4 ) ) ; if ( ( ( ( southWestLatitude != null ) && ( southWestLongitude != null ) ) && ( northEastLatitude != null ) ) && ( northEastLongitude != null ) ) { double latitude1 = java . lang . Double . parseDouble ( southWestLatitude ) ; double longitude1 = java . lang . Double . parseDouble ( southWestLongitude ) ; double latitude2 = java . lang . Double . parseDouble ( northEastLatitude ) ; double longitude2 = java . lang . Double . parseDouble ( northEastLongitude ) ; boundingBox = new double [ ] [ ] { new double [ ] { longitude1 , latitude1 } , new double [ ] { longitude2 , latitude2 } } ; com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( ( ( ( ( ( ( ( ( "Coordinates:<sp>'" + ( boundingBox [ 0 ] [ 0 ] ) ) + "maxBatchSize" 2 ) + ( boundingBox [ 0 ] [ 1 ] ) ) + "maxBatchSize" 2 ) + ( boundingBox [ 1 ] [ 0 ] ) ) + "maxBatchSize" 2 ) + ( boundingBox [ 1 ] [ 1 ] ) ) + "maxBatchSize" 4 ) ) ; haveFilters = true ; haveCoordinateFilter = true ; } if ( keywords != null ) { if ( ( keywords . trim ( ) . length ( ) ) != 0 ) { splitKeywords = keywords . split ( "consumerSecret" 2 ) ; for ( int i = 0 ; i < ( splitKeywords . length ) ; i ++ ) { splitKeywords [ i ] = splitKeywords [ i ] . trim ( ) ; } com . telefonica . iot . cygnus . sources . TwitterSource . LOGGER . info ( "consumerSecret" 0 , java . util . Arrays . toString ( splitKeywords ) ) ; haveFilters = true ; haveKeywordFilter = true ; } } maxBatchSize = context . getInteger ( "maxBatchSize" , maxBatchSize ) ; maxBatchDurationMillis = context . getInteger ( "maxBatchSize" 9 , maxBatchDurationMillis ) ; twitter4j . conf . ConfigurationBuilder cb = new twitter4j . conf . ConfigurationBuilder ( ) ; cb . setDebugEnabled ( true ) ; cb . setOAuthConsumerKey ( consumerKey ) ; cb . setOAuthConsumerSecret ( consumerSecret ) ; cb . setOAuthAccessToken ( accessToken ) ; cb . setOAuthAccessTokenSecret ( accessTokenSecret ) ; cb . setJSONStoreEnabled ( true ) ; twitterStream = new com . telefonica . iot . cygnus . sources . TwitterStreamFactory ( cb . build ( ) ) . getInstance ( ) ; }
org . junit . Assert . assertTrue ( true )
testQuery_ATTENDEE ( ) { try { com . github . caldav4j . util . GenerateQueryTest . log . info ( "VEVENT<sp>+<sp>ATTENDEE:" ) ; java . util . List < java . lang . String > a = new java . util . ArrayList ( ) ; a . add ( "ATTENDEE==mailto:lisa@example.com" ) ; com . github . caldav4j . util . GenerateQuery gq = new com . github . caldav4j . util . GenerateQuery ( ) ; gq . setFilter ( "VEVENT" , a ) ; com . github . caldav4j . util . GenerateQueryTest . log . info ( ( "setFilter()" + ( printQuery ( gq . generate ( ) ) ) ) ) ; gq = new com . github . caldav4j . util . GenerateQuery ( null , "VEVENT<sp>:<sp>ATTENDEE==mailto:lisa@example.com" ) ; com . github . caldav4j . util . GenerateQueryTest . log . info ( ( "Constructor:" + ( printQuery ( gq . generate ( ) ) ) ) ) ; } catch ( com . github . caldav4j . exceptions . CalDAV4JException e ) { e . printStackTrace ( ) ; "<AssertPlaceHolder>" ; } } generate ( ) { com . github . caldav4j . util . CalendarQuery query = new com . github . caldav4j . util . CalendarQuery ( ) ; query . addProperty ( CalDAVConstants . DNAME_GETETAG ) ; if ( allProp ) { query . addProperty ( CalDAVConstants . DNAME_ALLPROP ) ; } if ( ! ( noCalendarData ) ) { com . github . caldav4j . util . CalendarData calendarData = new com . github . caldav4j . util . CalendarData ( ) ; if ( ( ( recurrenceSetEnd ) != null ) || ( ( recurrenceSetStart ) != null ) ) { calendarData . setExpandOrLimitRecurrenceSet ( expandOrLimit ) ; calendarData . setRecurrenceSetStart ( recurrenceSetStart ) ; calendarData . setRecurrenceSetEnd ( recurrenceSetEnd ) ; } calendarData . setComp ( getComp ( ) ) ; query . setCalendarDataProp ( calendarData ) ; } else { if ( ( ( this . recurrenceSetEnd ) != null ) || ( ( this . recurrenceSetStart ) != null ) ) { throw new com . github . caldav4j . exceptions . CalDAV4JProtocolException ( "Bad<sp>query:<sp>you<sp>set<sp>noCalendarData<sp>but<sp>you<sp>have<sp>limit-recurrence-set" ) ; } } query . setCompFilter ( getFilter ( ) ) ; query . validate ( ) ; return query ; }
org . junit . Assert . assertTrue ( false )
contentsEqual ( ) { java . util . ArrayList < parser . Terminal > t1 = new java . util . ArrayList < parser . Terminal > ( ) ; java . util . ArrayList < parser . Terminal > t2 = new java . util . ArrayList < parser . Terminal > ( ) ; parser . Terminal t = new parser . Terminal ( "test" , null ) ; t1 . add ( t ) ; t2 . add ( t ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ! ( o instanceof parser . Item ) ) return false ; parser . Item other = ( ( parser . Item ) ( o ) ) ; if ( ( production . equals ( other . production ) ) && ( ( dot ) == ( other . dot ) ) ) { return ( itemLookahead ) == null ? ( other . itemLookahead ) == null : itemLookahead . equals ( other . itemLookahead ) ; } else { return false ; } }
org . junit . Assert . assertTrue ( t1 . equals ( t2 ) )
testConsistency ( ) { final com . yoghurt . crypto . transactions . shared . domain . Transaction tx = com . yoghurt . crypto . transactions . client . util . transaction . TransactionParseUtil . parseTransactionBytes ( com . googlecode . gwt . crypto . bouncycastle . util . encoders . Hex . decode ( com . yoghurt . crypto . transactions . client . util . transaction . TransactionEncodeUtilTest . TEST_TX ) ) ; final com . yoghurt . crypto . transactions . shared . domain . RawTransactionContainer reencoded = com . yoghurt . crypto . transactions . client . util . transaction . TransactionEncodeUtil . encodeTransaction ( tx ) ; final java . util . ArrayList < java . lang . Byte > bytes = new java . util . ArrayList < java . lang . Byte > ( ) ; for ( final java . util . Map . Entry < com . yoghurt . crypto . transactions . shared . domain . TransactionPartType , byte [ ] > part : reencoded ) { for ( final byte bite : part . getValue ( ) ) { bytes . add ( bite ) ; } } final byte [ ] primitiveBytes = new byte [ bytes . size ( ) ] ; int i = 0 ; for ( final byte bite : bytes ) { primitiveBytes [ ( i ++ ) ] = bite ; } final byte [ ] encode = com . googlecode . gwt . crypto . bouncycastle . util . encoders . Hex . encode ( primitiveBytes ) ; "<AssertPlaceHolder>" ; } encode ( byte [ ] ) { if ( ( input . length ) == 0 ) { return "" ; } input = com . yoghurt . crypto . transactions . client . util . address . Base58 . copyOfRange ( input , 0 , input . length ) ; int zeroCount = 0 ; while ( ( zeroCount < ( input . length ) ) && ( ( input [ zeroCount ] ) == 0 ) ) { ++ zeroCount ; } final byte [ ] temp = new byte [ ( input . length ) * 2 ] ; int j = temp . length ; int startAt = zeroCount ; while ( startAt < ( input . length ) ) { final byte mod = com . yoghurt . crypto . transactions . client . util . address . Base58 . divmod58 ( input , startAt ) ; if ( ( input [ startAt ] ) == 0 ) { ++ startAt ; } temp [ ( -- j ) ] = ( ( byte ) ( com . yoghurt . crypto . transactions . client . util . address . Base58 . ALPHABET [ mod ] ) ) ; } while ( ( j < ( temp . length ) ) && ( ( temp [ j ] ) == ( com . yoghurt . crypto . transactions . client . util . address . Base58 . ALPHABET [ 0 ] ) ) ) { ++ j ; } while ( ( -- zeroCount ) >= 0 ) { temp [ ( -- j ) ] = ( ( byte ) ( com . yoghurt . crypto . transactions . client . util . address . Base58 . ALPHABET [ 0 ] ) ) ; } final byte [ ] output = com . yoghurt . crypto . transactions . client . util . address . Base58 . copyOfRange ( temp , j , temp . length ) ; return com . yoghurt . crypto . transactions . client . util . TextConversionUtil . fromASCIIBytes ( output ) ; }
org . junit . Assert . assertEquals ( com . yoghurt . crypto . transactions . client . util . transaction . TransactionEncodeUtilTest . TEST_TX , new java . lang . String ( encode ) )
testReadRangeFromFileWithSplitsFromStart ( ) { org . apache . beam . sdk . options . PipelineOptions options = org . apache . beam . sdk . options . PipelineOptionsFactory . create ( ) ; java . lang . String header = "<h>" ; java . util . List < java . lang . String > data = new java . util . ArrayList ( ) ; for ( int i = 0 ; i < 10 ; i ++ ) { data . add ( header ) ; data . addAll ( createStringDataset ( 3 , 9 ) ) ; } java . lang . String fileName = "file" ; java . io . File file = createFileWithData ( fileName , data ) ; org . apache . beam . sdk . io . fs . MatchResult . Metadata metadata = org . apache . beam . sdk . io . FileSystems . matchSingleFileSpec ( file . getPath ( ) ) ; org . apache . beam . sdk . io . FileBasedSourceTest . TestFileBasedSource source1 = new org . apache . beam . sdk . io . FileBasedSourceTest . TestFileBasedSource ( metadata , 64 , 0 , 60 , header ) ; org . apache . beam . sdk . io . FileBasedSourceTest . TestFileBasedSource source2 = new org . apache . beam . sdk . io . FileBasedSourceTest . TestFileBasedSource ( metadata , 64 , 60 , Long . MAX_VALUE , header ) ; java . util . List < java . lang . String > expectedResults = new java . util . ArrayList ( ) ; expectedResults . addAll ( data ) ; expectedResults . removeAll ( java . util . Arrays . asList ( header ) ) ; java . util . List < java . lang . String > results = new java . util . ArrayList ( ) ; results . addAll ( readFromSource ( source1 , options ) ) ; results . addAll ( readFromSource ( source2 , options ) ) ; "<AssertPlaceHolder>" ; } containsInAnyOrder ( java . util . Collection ) { @ org . apache . beam . sdk . testing . SuppressWarnings ( { "rawtypes" , "unchecked" } ) final java . util . Collection < org . hamcrest . Matcher < ? super T > > matchers = ( ( java . util . Collection ) ( serializableMatchers ) ) ; return org . apache . beam . sdk . testing . SerializableMatchers . fromSupplier ( ( ) -> org . hamcrest . Matchers . containsInAnyOrder ( matchers ) ) ; }
org . junit . Assert . assertThat ( expectedResults , org . hamcrest . Matchers . containsInAnyOrder ( results . toArray ( ) ) )
testSerde ( ) { org . apache . druid . discovery . DruidService expected = new org . apache . druid . discovery . WorkerNodeService ( "1.1.1.1" , 100 , "v1" ) ; com . fasterxml . jackson . databind . ObjectMapper mapper = org . apache . druid . segment . TestHelper . makeJsonMapper ( ) ; org . apache . druid . discovery . DruidService actual = mapper . readValue ( mapper . writeValueAsString ( expected ) , org . apache . druid . discovery . DruidService . class ) ; "<AssertPlaceHolder>" ; } makeJsonMapper ( ) { final com . fasterxml . jackson . databind . ObjectMapper mapper = new org . apache . druid . jackson . DefaultObjectMapper ( ) ; mapper . setInjectableValues ( new com . fasterxml . jackson . databind . InjectableValues . Std ( ) . addValue ( org . apache . druid . math . expr . ExprMacroTable . class . getName ( ) , TestExprMacroTable . INSTANCE ) . addValue ( com . fasterxml . jackson . databind . ObjectMapper . class . getName ( ) , mapper ) . addValue ( DataSegment . PruneLoadSpecHolder . class , DataSegment . PruneLoadSpecHolder . DEFAULT ) ) ; return mapper ; }
org . junit . Assert . assertEquals ( expected , actual )
testSeekEventLoc_null ( ) { org . eclipse . tracecompass . tmf . ctf . core . context . CtfLocation loc = null ; fixture . seekEvent ( loc ) ; "<AssertPlaceHolder>" ; } seekEvent ( double ) { if ( ( fFile ) == null ) { return new org . eclipse . tracecompass . tmf . core . trace . TmfContext ( org . eclipse . tracecompass . btf . core . trace . BtfTrace . NULL_LOCATION , org . eclipse . tracecompass . tmf . core . trace . ITmfContext . UNKNOWN_RANK ) ; } try { long pos = ( java . lang . Math . round ( ( ratio * ( fFile . length ( ) ) ) ) ) - ( fDataOffset ) ; while ( pos > 0 ) { fFileInput . seek ( ( pos - 1 ) ) ; if ( ( fFileInput . read ( ) ) == '\n' ) { break ; } pos -- ; } final org . eclipse . tracecompass . tmf . core . trace . location . ITmfLocation location = new org . eclipse . tracecompass . tmf . core . trace . location . TmfLongLocation ( pos ) ; final org . eclipse . tracecompass . tmf . core . trace . ITmfContext context = seekEvent ( location ) ; context . setRank ( ITmfContext . UNKNOWN_RANK ) ; return context ; } catch ( final java . io . IOException e ) { org . eclipse . tracecompass . btf . core . Activator . logError ( ( "Error<sp>seeking<sp>event.<sp>File:<sp>" + ( getPath ( ) ) ) , e ) ; return new org . eclipse . tracecompass . tmf . core . parsers . custom . CustomTxtTraceContext ( org . eclipse . tracecompass . btf . core . trace . BtfTrace . NULL_LOCATION , org . eclipse . tracecompass . tmf . core . trace . ITmfContext . UNKNOWN_RANK ) ; } }
org . junit . Assert . assertNotNull ( fixture )
getBillingRunStartTimeInMs_eod ( ) { long expectedStartTime = ( org . oscm . configurationservice . bean . BillingOffsetTest . DAY_IN_MS ) - 1 ; long billingOffset = ( 4 * ( org . oscm . configurationservice . bean . BillingOffsetTest . DAY_IN_MS ) ) + expectedStartTime ; doReturn ( java . lang . Long . valueOf ( billingOffset ) ) . when ( configServiceBean ) . getLongConfigurationSetting ( ConfigurationKey . TIMER_INTERVAL_BILLING_OFFSET , Configuration . GLOBAL_CONTEXT ) ; long result = configServiceBean . getBillingRunStartTimeInMs ( ) ; "<AssertPlaceHolder>" ; } getBillingRunStartTimeInMs ( ) { throw new java . lang . UnsupportedOperationException ( ) ; }
org . junit . Assert . assertEquals ( expectedStartTime , result )
testSpdyRstStreamFrame ( ) { short type = 3 ; byte flags = 0 ; int length = 8 ; int streamId = ( ( io . netty . handler . codec . spdy . SpdyFrameDecoderTest . RANDOM . nextInt ( ) ) & 2147483647 ) | 1 ; int statusCode = ( io . netty . handler . codec . spdy . SpdyFrameDecoderTest . RANDOM . nextInt ( ) ) | 1 ; io . netty . buffer . ByteBuf buf = io . netty . buffer . Unpooled . buffer ( ( ( io . netty . handler . codec . spdy . SpdyCodecUtil . SPDY_HEADER_SIZE ) + length ) ) ; io . netty . handler . codec . spdy . SpdyFrameDecoderTest . encodeControlFrameHeader ( buf , type , flags , length ) ; buf . writeInt ( streamId ) ; buf . writeInt ( statusCode ) ; decoder . decode ( buf ) ; verify ( delegate ) . readRstStreamFrame ( streamId , statusCode ) ; "<AssertPlaceHolder>" ; buf . release ( ) ; } isReadable ( ) { return ( writerIndex ) > ( readerIndex ) ; }
org . junit . Assert . assertFalse ( buf . isReadable ( ) )
checkGetFirstDiagramNodeWithEmptyGraph ( ) { final org . kie . workbench . common . stunner . core . graph . Graph graph = new org . kie . workbench . common . stunner . core . graph . impl . GraphImpl ( "uuid" , new org . kie . workbench . common . stunner . core . graph . store . GraphNodeStoreImpl ( ) ) ; final org . kie . workbench . common . stunner . core . graph . Node < org . kie . workbench . common . stunner . core . graph . content . definition . Definition < org . kie . workbench . common . stunner . cm . definition . CaseManagementDiagram > , ? > fNode = org . kie . workbench . common . stunner . cm . util . CaseManagementUtils . getFirstDiagramNode ( graph ) ; "<AssertPlaceHolder>" ; } getFirstDiagramNode ( org . kie . workbench . common . stunner . core . graph . Graph ) { return org . kie . workbench . common . stunner . core . graph . util . GraphUtils . getFirstNode ( graph , org . kie . workbench . common . stunner . cm . definition . CaseManagementDiagram . class ) ; }
org . junit . Assert . assertNull ( fNode )
simpleBlockingPoll ( ) { org . ebayopensource . turmeric . runtime . sif . impl . transport . http . HTTPSyncAsyncClientTransportPoller completionQueue = new org . ebayopensource . turmeric . runtime . sif . impl . transport . http . HTTPSyncAsyncClientTransportPoller ( ) ; com . ebay . kernel . service . invocation . client . http . Request r1 = new com . ebay . kernel . service . invocation . client . http . Request ( "http://www.google.com" ) ; org . ebayopensource . turmeric . runtime . tests . common . sif . transport . HTTPSyncAsyncClientTransportPollerTest . client . send ( r1 , completionQueue ) ; java . util . List < java . util . concurrent . Future < ? > > futures = completionQueue . poll ( true ) ; "<AssertPlaceHolder>" ; } poll ( long ) { return poll ( timeout , TimeUnit . MILLISECONDS ) ; }
org . junit . Assert . assertTrue ( ( ( futures . size ( ) ) == 1 ) )
shouldReturnAbsentIfValueIsAbsent ( ) { when ( valueProvider . getAnnotationValue ( dataFactory ) ) . thenReturn ( java . util . Optional . < org . semanticweb . owlapi . model . OWLAnnotationValue > empty ( ) ) ; "<AssertPlaceHolder>" ; } getAnnotation ( org . protege . editor . owl . OWLEditorKit ) { return java . util . Optional . empty ( ) ; }
org . junit . Assert . assertThat ( provider . getAnnotation ( dataFactory ) , org . hamcrest . core . Is . is ( java . util . Optional . empty ( ) ) )
testReadWriteObjectGenerative ( ) { com . liveramp . generative . Generative . runTests ( 10 , ( testNum , gen ) -> { List < java . lang . String > path = gen . namedVar ( "path" ) . listOfLength ( gen . anyStringOfLengthUpTo ( 1000 ) , gen . anyBoundedInteger ( 0 , 20 ) ) . get ( ) ; HashMap < java . lang . String , java . lang . Integer > map = gen . anyStringOfLengthUpTo ( 1000 ) . stream ( ) . map ( ( s ) -> org . apache . commons . lang3 . tuple . Pair . of ( s , gen . anyPositiveInteger ( ) . get ( ) ) ) . limit ( gen . anyBoundedInteger ( 0 , 5000 ) . get ( ) ) . collect ( com . liveramp . java_support . functional . Fns8 . pairsToMap ( ) ) ; Stream < Pair < java . lang . String , java . lang . Integer > > dataStream = map . entrySet ( ) . stream ( ) . map ( ( e ) -> org . apache . commons . lang3 . tuple . Pair . of ( e . getKey ( ) , e . getValue ( ) ) ) ; org . apache . hadoop . fs . FileSystem fs = com . liveramp . cascading_ext . FileSystemHelper . getFileSystemForPath ( getTestRoot ( ) ) ; com . google . gson . Gson gson = new com . google . gson . Gson ( ) ; org . apache . hadoop . fs . Path dataPath = new org . apache . hadoop . fs . Path ( getTestRootPath ( ) , "data" ) ; com . rapleaf . cascading_ext . hdfs_utils . HdfsGsonHelper . writeFromStream ( fs , gson , dataPath , dataStream , . class , path . toArray ( new java . lang . String [ 0 ] ) ) ; HashMap < java . lang . String , java . lang . Integer > result = com . rapleaf . cascading_ext . hdfs_utils . HdfsGsonHelper . streamObject ( fs , gson , dataPath , . class , path . toArray ( new java . lang . String [ 0 ] ) ) . collect ( com . liveramp . java_support . functional . Fns8 . pairsToMap ( ) ) ; "<AssertPlaceHolder>" ; } ) ; } collect ( cascading . tuple . TupleEntry ) { tupleEntryCollector . add ( tupleEntry . getTuple ( ) ) ; }
org . junit . Assert . assertEquals ( map , result )
asInt ( ) { "<AssertPlaceHolder>" ; } asInt ( ) { org . junit . Assert . assertEquals ( 23 , new org . eclipse . rap . json . JsonNumber ( "23" ) . asInt ( ) ) ; }
org . junit . Assert . assertEquals ( 23 , new org . eclipse . rap . json . JsonNumber ( "23" ) . asInt ( ) )
testValidNamesTag ( ) { final java . lang . String tagSrc = "<cfcomponent>\r\n" + ( ( ( ( "<cffunction<sp>name=\"test\">\r\n" + "\t<cfargument<sp>name=\"first_name\">\r\n" ) + "\t<cfargument<sp>name=\"firstname\">\r\n" ) + "</cffunction>\r\n" ) + "</cfcomponent>" ) ; com . cflint . api . CFLintResult lintresult = cfBugs . scan ( tagSrc , "test" ) ; java . util . Collection < java . util . List < com . cflint . BugInfo > > result = lintresult . getIssues ( ) . values ( ) ; "<AssertPlaceHolder>" ; } size ( ) { int size = 0 ; for ( final java . util . List < ? > list : bugs . values ( ) ) { size += list . size ( ) ; } return size ; }
org . junit . Assert . assertEquals ( 0 , result . size ( ) )
test_Negativity_In_Equality_For_IOU ( ) { com . ripple . core . coretypes . Amount a = com . ripple . core . AmountTest . amt ( "1/USD/rNDKeo9RrCiRdfsMG8AdoZvNZxHASGzbZL" ) ; com . ripple . core . coretypes . Amount b = com . ripple . core . AmountTest . amt ( "-1/USD/rNDKeo9RrCiRdfsMG8AdoZvNZxHASGzbZL" ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ! ( o instanceof org . ripple . bouncycastle . asn1 . eac . PackedDate ) ) { return false ; } org . ripple . bouncycastle . asn1 . eac . PackedDate other = ( ( org . ripple . bouncycastle . asn1 . eac . PackedDate ) ( o ) ) ; return org . ripple . bouncycastle . util . Arrays . areEqual ( time , other . time ) ; }
org . junit . Assert . assertFalse ( a . equals ( b ) )
testIsValidNameNull ( ) { "<AssertPlaceHolder>" ; } isValidName ( java . lang . String ) { return com . liferay . document . library . kernel . util . DLValidatorUtil . _dlValidator . isValidName ( name ) ; }
org . junit . Assert . assertFalse ( com . liferay . document . library . kernel . util . DLValidatorUtil . isValidName ( null ) )
testResetUserPassword ( ) { usersPage . getOtherUser ( 1L ) ; usersPage . sendPasswordReset ( ) ; "<AssertPlaceHolder>" ; } checkSuccessNotification ( ) { return pageUtilities . checkSuccessNotification ( ) ; }
org . junit . Assert . assertTrue ( usersPage . checkSuccessNotification ( ) )
nameFirstUpdateIsAnyUpdate ( ) { final org . libdohj . script . NameScript ns = getNameFirstUpdateNameScript ( ) ; "<AssertPlaceHolder>" ; } isAnyUpdate ( ) { switch ( op ) { case org . libdohj . script . NameScript . OP_NAME_NEW : return false ; case org . libdohj . script . NameScript . OP_NAME_FIRSTUPDATE : case org . libdohj . script . NameScript . OP_NAME_UPDATE : return true ; default : throw new org . bitcoinj . core . ScriptException ( "Not<sp>a<sp>name<sp>op" ) ; } }
org . junit . Assert . assertTrue ( ns . isAnyUpdate ( ) )
getCounterManage_test ( ) { scouter . server . CounterManager counterManager = scouter . server . CounterManager . getInstance ( ) ; "<AssertPlaceHolder>" ; } getInstance ( ) { if ( ( scouter . server . CounterManager . instance ) == null ) { synchronized ( scouter . server . CounterManager . class ) { if ( ( scouter . server . CounterManager . instance ) == null ) { scouter . server . CounterManager . instance = new scouter . server . CounterManager ( ) ; } } } return scouter . server . CounterManager . instance ; }
org . junit . Assert . assertNotNull ( counterManager )
decodeExtended ( ) { final java . io . ByteArrayInputStream stream = new java . io . ByteArrayInputStream ( extended ) ; final com . flagstone . transform . coder . SWFDecoder decoder = new com . flagstone . transform . coder . SWFDecoder ( stream ) ; fixture = new com . flagstone . transform . video . DefineVideo ( decoder ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( fixture )
stream_of_strings_to_ints_array ( ) { java . lang . Integer [ ] stringArray = java . util . stream . Stream . of ( 1 , 2 , 3 ) . toArray ( java . lang . Integer [ ] :: new ) ; com . levelup . java . util . stream . ConvertStreamToArray . logger . info ( java . util . Arrays . toString ( stringArray ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertTrue ( ( ( stringArray . length ) == 3 ) )
testReadProjectWithWorkspaceThatDoesNotExist ( ) { org . eclipse . orion . server . core . metastore . IMetaStore metaStore = org . eclipse . orion . server . core . OrionConfiguration . getMetaStore ( ) ; org . eclipse . orion . server . core . metastore . UserInfo userInfo = new org . eclipse . orion . server . core . metastore . UserInfo ( ) ; userInfo . setUserName ( testUserLogin ) ; userInfo . setFullName ( testUserLogin ) ; metaStore . createUser ( userInfo ) ; java . lang . String workspaceName = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStore . DEFAULT_WORKSPACE_NAME ; java . lang . String workspaceId = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . encodeWorkspaceId ( userInfo . getUniqueId ( ) , workspaceName ) ; org . eclipse . orion . server . core . metastore . ProjectInfo readProjectInfo = metaStore . readProject ( workspaceId , "Project<sp>Zero" ) ; "<AssertPlaceHolder>" ; } readProject ( java . lang . String , java . lang . String ) { java . lang . String userId = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . decodeUserIdFromWorkspaceId ( workspaceId ) ; if ( userId == null ) { if ( logger . isDebugEnabled ( ) ) { logger . debug ( ( "SimpleMetaStore.readProject:<sp>requested<sp>with<sp>a<sp>bad<sp>userId<sp>in<sp>the<sp>workspaceId<sp>" + workspaceId ) ) ; } return null ; } org . eclipse . orion . server . core . resources . FileLocker . Lock lock = null ; try { lock = getUserLock ( userId ) . lock ( true ) ; java . lang . String encodedWorkspaceName = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . decodeWorkspaceNameFromWorkspaceId ( workspaceId ) ; if ( encodedWorkspaceName == null ) { if ( logger . isDebugEnabled ( ) ) { logger . debug ( ( "SimpleMetaStore.readProject:<sp>requested<sp>with<sp>a<sp>bad<sp>workspaceId<sp>" + workspaceId ) ) ; } return null ; } java . io . File userMetaFolder = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . readMetaUserFolder ( getRootLocation ( ) , userId ) ; java . io . File workspaceMetaFolder = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . readMetaFolder ( userMetaFolder , encodedWorkspaceName ) ; if ( workspaceMetaFolder == null ) { if ( logger . isDebugEnabled ( ) ) { logger . debug ( ( "SimpleMetaStore.readProject:<sp>workspaceMetaFolder<sp>does<sp>not<sp>exist<sp>for<sp>workspace<sp>" + workspaceId ) ) ; } return null ; } java . lang . String projectId = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . encodeProjectIdFromProjectName ( projectName ) ; java . lang . String metafileId = projectId ; if ( ! ( org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . encodeWorkspaceId ( userId , org . eclipse . orion . internal . server . core . metastore . SimpleMetaStore . DEFAULT_WORKSPACE_NAME ) . equals ( workspaceId ) ) ) { metafileId = ( workspaceId + ( SimpleMetaStoreUtil . SEPARATOR ) ) + metafileId ; } org . json . JSONObject jsonObject = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . readMetaFile ( userMetaFolder , metafileId ) ; org . eclipse . orion . server . core . metastore . ProjectInfo projectInfo = new org . eclipse . orion . server . core . metastore . ProjectInfo ( ) ; if ( jsonObject == null ) { if ( ( org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . isMetaFolder ( workspaceMetaFolder , projectId ) ) && ( ! ( org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . isMetaFile ( userMetaFolder , metafileId ) ) ) ) { java . io . File projectMetaFolder = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . readMetaFolder ( workspaceMetaFolder , projectId ) ; if ( logger . isDebugEnabled ( ) ) { logger . info ( ( ( ( "SimpleMetaStore.readProject:<sp>the<sp>project<sp>folder<sp>" + ( projectMetaFolder . toString ( ) ) ) + "<sp>exists<sp>but<sp>the<sp>project<sp>json<sp>file<sp>does<sp>not,<sp>so<sp>creating<sp>it<sp>in<sp>" ) + workspaceId ) ) ; } java . net . URI projectLocation = projectMetaFolder . toURI ( ) ; projectInfo . setFullName ( projectName ) ; projectInfo . setWorkspaceId ( workspaceId ) ; projectInfo . setContentLocation ( projectLocation ) ; lock . release ( ) ; createProject ( projectInfo ) ; lock = getUserLock ( userId ) . lock ( true ) ; jsonObject = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . readMetaFile ( userMetaFolder , metafileId ) ; } else { return null ; } } try { projectInfo . setUniqueId ( jsonObject . getString ( MetadataInfo . UNIQUE_ID ) ) ; projectInfo . setWorkspaceId ( jsonObject . getString ( "WorkspaceId" ) ) ; projectInfo . setFullName ( jsonObject . getString ( UserConstants . FULL_NAME ) ) ; if ( jsonObject . has ( "ContentLocation" ) ) { java . lang . String decodedContentLocation = org . eclipse . orion . internal . server . core . metastore . SimpleMetaStoreUtil . decodeProjectContentLocation ( jsonObject . getString ( "ContentLocation" ) ) ; projectInfo . setContentLocation ( new java . net . URI ( decodedContentLocation ) ) ; } setProperties ( projectInfo , jsonObject . getJSONObject ( "Properties" ) ) ; projectInfo . flush ( ) ; } catch ( org . json . JSONException e ) { throw new org . eclipse . core . runtime . CoreException ( new org . eclipse . core . runtime . Status ( org . eclipse . core . runtime . IStatus . ERROR , org . eclipse . orion . server . core . ServerConstants . PI_SERVER_CORE , 1 , (
org . junit . Assert . assertNull ( readProjectInfo )
testGetResultForNode ( ) { result = new ch . puzzle . itc . mobiliar . presentation . generator . ComparedGenerationResult ( originalResult ) ; ch . puzzle . itc . mobiliar . business . generator . control . NodeGenerationResult newResult = result . getResultForNode ( nodeResult . getNode ( ) , originalResult ) ; "<AssertPlaceHolder>" ; } getNode ( ) { return node ; }
org . junit . Assert . assertTrue ( ( newResult == ( this . nodeResult ) ) )
testGetParameters ( ) { java . lang . String name = "Extract<sp>Me<sp>All" ; java . lang . String abbrName = "EML" ; org . lnu . is . domain . reason . Reason entity = new org . lnu . is . domain . reason . Reason ( ) ; entity . setName ( name ) ; entity . setAbbrName ( abbrName ) ; java . util . Map < java . lang . String , java . lang . Object > expected = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; expected . put ( "name" , name ) ; expected . put ( "abbrName" , abbrName ) ; expected . put ( "status" , RowStatus . ACTIVE ) ; expected . put ( "userGroups" , groups ) ; java . util . Map < java . lang . String , java . lang . Object > actual = unit . getParameters ( entity ) ; "<AssertPlaceHolder>" ; } getParameters ( org . springframework . web . context . request . NativeWebRequest ) { java . util . Map < java . lang . String , java . lang . Object > resultMap = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; java . util . Map < java . lang . String , java . lang . String > pathVariables = ( ( java . util . Map < java . lang . String , java . lang . String > ) ( webRequest . getAttribute ( HandlerMapping . URI_TEMPLATE_VARIABLES_ATTRIBUTE , RequestAttributes . SCOPE_REQUEST ) ) ) ; java . util . Map < java . lang . String , java . lang . Object > requestParams = getRequestParameterMap ( webRequest ) ; for ( Map . Entry < java . lang . String , java . lang . Object > entry : requestParams . entrySet ( ) ) { resultMap . put ( entry . getKey ( ) , entry . getValue ( ) ) ; } resultMap . putAll ( pathVariables ) ; return resultMap ; }
org . junit . Assert . assertEquals ( expected , actual )
testExpression70 ( ) { net . objecthunter . exp4j . Expression e = new net . objecthunter . exp4j . ExpressionBuilder ( "2xx" ) . variables ( "x" ) . build ( ) . setVariable ( "x" , Math . E ) ; "<AssertPlaceHolder>" ; } evaluate ( ) { final net . objecthunter . exp4j . ArrayStack output = new net . objecthunter . exp4j . ArrayStack ( ) ; for ( int i = 0 ; i < ( tokens . length ) ; i ++ ) { net . objecthunter . exp4j . tokenizer . Token t = tokens [ i ] ; else if ( ( t . getType ( ) ) == ( net . objecthunter . exp4j . tokenizer . Token . TOKEN_VARIABLE ) ) { final java . lang . String name = ( ( net . objecthunter . exp4j . tokenizer . VariableToken ) ( t ) ) . getName ( ) ; final java . lang . Double value = this . variables . get ( name ) ; if ( value == null ) { throw new java . lang . IllegalArgumentException ( ( ( "No<sp>value<sp>has<sp>been<sp>set<sp>for<sp>the<sp>setVariable<sp>'" + name ) + "'." ) ) ; } output . push ( value ) ; } else if ( ( t . getType ( ) ) == ( net . objecthunter . exp4j . tokenizer . Token . TOKEN_OPERATOR ) ) { net . objecthunter . exp4j . tokenizer . OperatorToken op = ( ( net . objecthunter . exp4j . tokenizer . OperatorToken ) ( t ) ) ; if ( ( output . size ( ) ) < ( op . getOperator ( ) . getNumOperands ( ) ) ) { throw new java . lang . IllegalArgumentException ( ( ( "Invalid<sp>number<sp>of<sp>operands<sp>available<sp>for<sp>'" + ( op . getOperator ( ) . getSymbol ( ) ) ) + "'<sp>operator" ) ) ; } if ( ( op . getOperator ( ) . getNumOperands ( ) ) == 2 ) { double rightArg = output . pop ( ) ; double leftArg = output . pop ( ) ; output . push ( op . getOperator ( ) . apply ( leftArg , rightArg ) ) ; } else if ( ( op . getOperator ( ) . getNumOperands ( ) ) == 1 ) { double arg = output . pop ( ) ; output . push ( op . getOperator ( ) . apply ( arg ) ) ; } } else if ( ( t . getType ( ) ) == ( net . objecthunter . exp4j . tokenizer . Token . TOKEN_FUNCTION ) ) { net . objecthunter . exp4j . tokenizer . FunctionToken func = ( ( net . objecthunter . exp4j . tokenizer . FunctionToken ) ( t ) ) ; final int numArguments = func . getFunction ( ) . getNumArguments ( ) ; if ( ( output . size ( ) ) < numArguments ) { throw new java . lang . IllegalArgumentException ( ( ( "Invalid<sp>number<sp>of<sp>arguments<sp>available<sp>for<sp>'" + ( func . getFunction ( ) . getName ( ) ) ) + "'<sp>function" ) ) ; } double [ ] args = new double [ numArguments ] ; for ( int j = numArguments - 1 ; j >= 0 ; j -- ) { args [ j ] = output . pop ( ) ; } output . push ( func . getFunction ( ) . apply ( args ) ) ; } } if ( ( output . size ( ) ) > 1 ) { throw new java . lang . IllegalArgumentException ( "Invalid<sp>number<sp>of<sp>items<sp>on<sp>the<sp>output<sp>queue.<sp>Might<sp>be<sp>caused<sp>by<sp>an<sp>invalid<sp>number<sp>of<sp>arguments<sp>for<sp>a<sp>function." ) ; } return output . pop ( ) ; }
org . junit . Assert . assertEquals ( ( ( 2 * ( java . lang . Math . Math . E ) ) * ( java . lang . Math . Math . E ) ) , e . evaluate ( ) , 0.0 )
dropWhile_A$Function1_Nil ( ) { com . m3 . scalaflavor4j . Seq < java . lang . Integer > seq = com . m3 . scalaflavor4j . Seq . apply ( ) ; com . m3 . scalaflavor4j . Seq < java . lang . Integer > actual = seq . dropWhile ( new com . m3 . scalaflavor4j . F1 < java . lang . Integer , java . lang . Boolean > ( ) { public com . m3 . scalaflavor4j . Boolean apply ( java . lang . Integer v1 ) { return v1 >= 3 ; } } ) ; "<AssertPlaceHolder>" ; } size ( ) { return 0 ; }
org . junit . Assert . assertThat ( actual . size ( ) , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( 0 ) ) )
testShouldBeEqual ( ) { com . eclipsesource . tabris . ui . PageData pageData1 = new com . eclipsesource . tabris . ui . PageData ( ) ; pageData1 . set ( "a" , "1" ) ; com . eclipsesource . tabris . ui . PageData pageData2 = new com . eclipsesource . tabris . ui . PageData ( ) ; pageData2 . set ( "a" , "1" ) ; boolean equal = pageData1 . equals ( pageData2 ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) { return true ; } if ( obj == null ) { return false ; } if ( ( getClass ( ) ) != ( obj . getClass ( ) ) ) { return false ; } com . eclipsesource . tabris . passepartout . Bounds other = ( ( com . eclipsesource . tabris . passepartout . Bounds ) ( obj ) ) ; if ( ( height ) != ( other . height ) ) { return false ; } if ( ( width ) != ( other . width ) ) { return false ; } if ( ( x ) != ( other . x ) ) { return false ; } if ( ( y ) != ( other . y ) ) { return false ; } return true ; }
org . junit . Assert . assertTrue ( equal )
testUserExists ( ) { boolean b = dao . exists ( ( - 1L ) ) ; "<AssertPlaceHolder>" ; } exists ( PK extends java . io . Serializable ) { T entity = ( ( T ) ( super . getHibernateTemplate ( ) . get ( this . persistentClass , id ) ) ) ; return entity != null ; }
org . junit . Assert . assertTrue ( b )
testGetUpnpMimeType ( ) { org . seamless . util . MimeType mimeType = net . holmes . core . business . streaming . upnp . UpnpUtils . getUpnpMimeType ( net . holmes . core . business . mimetype . model . MimeType . valueOf ( "image/png" ) ) ; "<AssertPlaceHolder>" ; } valueOf ( java . lang . String ) { return mimeType == null ? null : new net . holmes . core . business . mimetype . model . MimeType ( mimeType ) ; }
org . junit . Assert . assertNotNull ( mimeType )
applyTest ( ) { java . lang . String [ ] [ ] tests = new java . lang . String [ ] [ ] { new java . lang . String [ ] { "{{<sp>'foo+bar'<sp>|<sp>url_decode<sp>}}" , "foo<sp>bar" } , new java . lang . String [ ] { "{{<sp>'foo%20bar'<sp>|<sp>url_decode<sp>}}" , "foo<sp>bar" } , new java . lang . String [ ] { "{{<sp>'foo%2B1%40example.com'<sp>|<sp>url_decode<sp>}}" , "foo+1@example.com" } , new java . lang . String [ ] { "{{<sp>nil<sp>|<sp>url_decode<sp>}}" , "" } } ; for ( java . lang . String [ ] test : tests ) { liqp . Template template = liqp . Template . parse ( test [ 0 ] ) ; java . lang . String rendered = java . lang . String . valueOf ( template . render ( ) ) ; "<AssertPlaceHolder>" ; } } render ( ) { return render ( new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ) ; }
org . junit . Assert . assertThat ( rendered , org . hamcrest . CoreMatchers . is ( test [ 1 ] ) )
testStripSequenceNumberBlank ( ) { "<AssertPlaceHolder>" ; } stripSequenceNumber ( java . lang . String ) { if ( ( isBlank ( name ) ) || ( ! ( java . lang . Character . isDigit ( name . charAt ( ( ( name . length ( ) ) - 1 ) ) ) ) ) ) { return name ; } return au . gov . ga . worldwind . animator . util . FileUtil . stripSequenceNumber ( name . substring ( 0 , ( ( name . length ( ) ) - 1 ) ) ) ; }
org . junit . Assert . assertEquals ( "<sp>" , au . gov . ga . worldwind . animator . util . FileUtil . stripSequenceNumber ( "<sp>" ) )
testRouterServicePathLookUp ( ) { java . lang . String path = "/v3/namespaces/default//apps/ResponseCodeAnalytics/services/LogAnalyticsService/status" ; io . netty . handler . codec . http . HttpRequest httpRequest = new io . netty . handler . codec . http . DefaultHttpRequest ( io . cdap . cdap . gateway . router . RouterPathLookupTest . VERSION , new io . netty . handler . codec . http . HttpMethod ( "GET" ) , path ) ; io . cdap . cdap . gateway . router . RouteDestination result = io . cdap . cdap . gateway . router . RouterPathLookupTest . pathLookup . getRoutingService ( path , httpRequest ) ; "<AssertPlaceHolder>" ; } getRoutingService ( java . lang . String , io . netty . handler . codec . http . HttpRequest ) { try { java . lang . String method = httpRequest . method ( ) . name ( ) ; io . cdap . cdap . gateway . router . RouterPathLookup . AllowedMethod requestMethod = io . cdap . cdap . gateway . router . RouterPathLookup . AllowedMethod . valueOf ( method ) ; java . lang . String [ ] uriParts = org . apache . commons . lang . StringUtils . split ( requestPath , '/' ) ; if ( uriParts [ 0 ] . equals ( Constants . Gateway . API_VERSION_3_TOKEN ) ) { return getV3RoutingService ( uriParts , requestMethod ) ; } } catch ( java . lang . Exception e ) { } return io . cdap . cdap . gateway . router . RouterPathLookup . APP_FABRIC_HTTP ; }
org . junit . Assert . assertEquals ( RouterPathLookup . APP_FABRIC_HTTP , result )
genericWithReflection ( ) { org . apache . avro . generic . GenericRecord savedRecord = new org . apache . avro . generic . GenericData . Record ( org . apache . crunch . test . Person . SCHEMA$ ) ; savedRecord . put ( "name" , "John<sp>Doe" ) ; savedRecord . put ( "age" , 42 ) ; savedRecord . put ( "siblingnames" , com . google . common . collect . Lists . newArrayList ( "Jimmy" , "Jane" ) ) ; populateGenericFile ( com . google . common . collect . Lists . newArrayList ( savedRecord ) , Person . SCHEMA . ) ; org . apache . crunch . Pipeline pipeline = new org . apache . crunch . impl . mr . MRPipeline ( org . apache . crunch . io . avro . AvroFileSourceTargetIT . class , tmpDir . getDefaultConfiguration ( ) ) ; org . apache . crunch . PCollection < org . apache . crunch . test . Person > genericCollection = pipeline . read ( org . apache . crunch . io . At . avroFile ( avroFile . getAbsolutePath ( ) , org . apache . crunch . types . avro . Avros . records ( org . apache . crunch . test . Person . class ) ) ) ; org . apache . crunch . PTable < java . lang . Long , org . apache . crunch . test . StringWrapper > pt = genericCollection . parallelDo ( new org . apache . crunch . MapFn < org . apache . crunch . test . Person , org . apache . crunch . Pair < java . lang . Long , org . apache . crunch . test . StringWrapper > > ( ) { @ org . apache . crunch . io . avro . Override public org . apache . crunch . Pair < java . lang . Long , org . apache . crunch . test . StringWrapper > map ( org . apache . crunch . test . Person input ) { return org . apache . crunch . Pair . of ( 1L , new org . apache . crunch . test . StringWrapper ( input . getName ( ) . toString ( ) ) ) ; } } , org . apache . crunch . types . avro . Avros . tableOf ( org . apache . crunch . types . avro . Avros . longs ( ) , org . apache . crunch . types . avro . Avros . reflects ( org . apache . crunch . test . StringWrapper . class ) ) ) . groupByKey ( ) . ungroup ( ) ; java . util . List < org . apache . crunch . Pair < java . lang . Long , org . apache . crunch . test . StringWrapper > > ret = com . google . common . collect . Lists . newArrayList ( pt . materialize ( ) ) ; pipeline . done ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return 3 ; }
org . junit . Assert . assertEquals ( 1 , ret . size ( ) )
testPass ( ) { "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( 1 , 1 )
compareListContentsFullOverlap ( ) { java . util . List < java . lang . String > names1 = new java . util . ArrayList ( ) ; java . util . List < java . lang . String > names2 = new java . util . ArrayList ( ) ; names1 . add ( "bob" ) ; names1 . add ( "alice" ) ; names2 . add ( "bob" ) ; names2 . add ( "alice" ) ; "<AssertPlaceHolder>" ; } checkContainsSame ( java . util . List , java . util . Collection ) { if ( ( list1 . size ( ) ) != ( list2 . size ( ) ) ) return false ; java . util . Collection < java . lang . Object > list3 = new java . util . ArrayList ( ) ; list3 . addAll ( list1 ) ; list3 . removeAll ( list2 ) ; return ( list3 . size ( ) ) == 0 ; }
org . junit . Assert . assertTrue ( checkContainsSame ( names1 , names2 ) )
testPolling ( ) { java . io . File tmpDir = tempDir . newFolder ( ) ; org . deeplearning4j . perf . listener . SystemPolling systemPolling = new org . deeplearning4j . perf . listener . SystemPolling . Builder ( ) . outputDirectory ( tmpDir ) . pollEveryMillis ( 1000 ) . build ( ) ; systemPolling . run ( ) ; java . lang . Thread . sleep ( 8000 ) ; systemPolling . stopPolling ( ) ; java . io . File [ ] files = tmpDir . listFiles ( ) ; "<AssertPlaceHolder>" ; java . lang . String yaml = org . apache . commons . io . FileUtils . readFileToString ( files [ 0 ] ) ; org . deeplearning4j . perf . listener . HardwareMetric fromYaml = org . deeplearning4j . perf . listener . HardwareMetric . fromYaml ( yaml ) ; System . out . println ( fromYaml ) ; } stopPolling ( ) { scheduledExecutorService . shutdownNow ( ) ; }
org . junit . Assert . assertTrue ( ( ( files != null ) && ( ( files . length ) > 0 ) ) )
testRandomByteArrayTransfer2 ( ) { byte [ ] value = new byte [ ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) * 2 ] ; for ( int i = 0 ; i < ( ( ( buffer . capacity ( ) ) - ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) ) + 1 ) ; i += org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) { random . nextBytes ( value ) ; buffer . setBytes ( i , value , random . nextInt ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) , org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) ; } random . setSeed ( seed ) ; byte [ ] expectedValueContent = new byte [ ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) * 2 ] ; org . apache . flink . shaded . netty4 . io . netty . buffer . ByteBuf expectedValue = wrappedBuffer ( expectedValueContent ) ; for ( int i = 0 ; i < ( ( ( buffer . capacity ( ) ) - ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) ) + 1 ) ; i += org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) { random . nextBytes ( expectedValueContent ) ; int valueOffset = random . nextInt ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) ; buffer . getBytes ( i , value , valueOffset , org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) ; for ( int j = valueOffset ; j < ( valueOffset + ( org . apache . flink . runtime . io . network . buffer . AbstractByteBufTest . BLOCK_SIZE ) ) ; j ++ ) { "<AssertPlaceHolder>" ; } } } getByte ( java . lang . String ) { addToDefaults ( key , null ) ; java . lang . String value = getRequired ( key ) ; return java . lang . Byte . valueOf ( value ) ; }
org . junit . Assert . assertEquals ( expectedValue . getByte ( j ) , value [ j ] )
testGetPropertyInstanceNoTargetType ( ) { org . finra . herd . model . dto . ConfigurationValue configurationValue = org . finra . herd . model . dto . ConfigurationValue . HERD_ENVIRONMENT ; java . lang . String value = configurationHelper . getProperty ( configurationValue ) ; "<AssertPlaceHolder>" ; } getDefaultValue ( ) { return defaultValue ; }
org . junit . Assert . assertEquals ( "value" , configurationValue . getDefaultValue ( ) , value )
testBadCast ( ) { org . eurekastreams . commons . test . EasyMatcher < java . lang . String > sut = new org . eurekastreams . commons . test . EasyMatcher < java . lang . String > ( ) { @ org . eurekastreams . commons . test . Override protected boolean isMatch ( final java . lang . String inTestObject ) { java . lang . Object o = "string" ; java . util . Date d = ( ( java . util . Date ) ( o ) ) ; return false ; } } ; "<AssertPlaceHolder>" ; } matches ( java . lang . Object ) { return org . eurekastreams . commons . test . IsEqualInternally . areEqualInternally ( item , compareAgainst ) ; }
org . junit . Assert . assertFalse ( sut . matches ( null ) )
testGetParent ( ) { org . eclipse . xtext . nodemodel . impl . AbstractNode node = createNode ( ) ; org . eclipse . xtext . nodemodel . ICompositeNode parent = builder . newRootNode ( "input" ) ; builder . addChild ( parent , node ) ; "<AssertPlaceHolder>" ; } getParent ( ) { if ( ( grammarElementIdx ) == 1 ) return delegate ; return new org . eclipse . xtext . nodemodel . impl . SyntheticCompositeNode ( delegate , ( ( grammarElementIdx ) - 1 ) ) ; }
org . junit . Assert . assertSame ( parent , node . getParent ( ) )
split_string_colon_guava ( ) { java . util . List < java . lang . String > elementsInString = com . google . common . base . Splitter . on ( ":" ) . splitToList ( "This:is:a:sentence:by:colon" ) ; com . levelup . java . string . SplitStringByColon . logger . info ( elementsInString ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( 6 , elementsInString . size ( ) )
testOpslaan ( ) { nl . bzk . brp . preview . model . Bericht bericht = new nl . bzk . brp . preview . model . Bericht ( ) ; bericht . setPartij ( "<sp>and<sp>indprevalidatie<sp>=<sp>true" 0 ) ; bericht . setBericht ( "qwe" ) ; bericht . setBerichtDetails ( "asd" ) ; bericht . setAantalMeldingen ( 2 ) ; java . util . Calendar verzondenOp = java . util . Calendar . getInstance ( ) ; verzondenOp . clear ( ) ; verzondenOp . set ( 2004 , Calendar . OCTOBER , 19 , 10 , 23 , 54 ) ; bericht . setVerzondenOp ( verzondenOp ) ; bericht . setBurgerZakenModule ( "zxc" ) ; bericht . setSoortBijhouding ( OndersteundeBijhoudingsTypes . GEBOORTE ) ; bericht . setPrevalidatie ( true ) ; berichtenDao . opslaan ( bericht ) ; java . lang . String sql = "<sp>and<sp>indprevalidatie<sp>=<sp>true" 2 + ( ( ( ( ( ( ( "<sp>and<sp>indprevalidatie<sp>=<sp>true" 1 + "<sp>and<sp>bericht<sp>=<sp>'qwe'" ) + "<sp>and<sp>berichtdetails<sp>=<sp>'asd'" ) + "<sp>and<sp>aantalmeldingen<sp>=<sp>2" ) + "<sp>and<sp>tsverzonden<sp>=<sp>TIMESTAMP<sp>'2004-10-19<sp>10:23:54'" ) + "<sp>and<sp>bzm<sp>=<sp>'zxc'" ) + "<sp>and<sp>soortactie<sp>=<sp>'GEBOORTE'" ) + "<sp>and<sp>indprevalidatie<sp>=<sp>true" ) ; long count = jdbcTemplate . queryForLong ( sql , ( ( java . util . Map < java . lang . String , ? > ) ( null ) ) ) ; "<AssertPlaceHolder>" ; } queryForLong ( java . lang . String , int , java . lang . Object [ ] ) { int count = 0 ; while ( count < maxCount ) { count ++ ; try { return jdbcTemplate . queryForLong ( sql , params ) ; } catch ( final org . springframework . dao . EmptyResultDataAccessException e ) { try { java . lang . Thread . sleep ( ( ( nl . moderniseringgba . migratie . test . isc . environment . ExternalTestEnvironment . TIMEOUT ) * count ) ) ; } catch ( final java . lang . InterruptedException e1 ) { } } } return null ; }
org . junit . Assert . assertEquals ( 1 , count )
testLocalAddress ( ) { java . net . InetSocketAddress localAddress = new java . net . InetSocketAddress ( "localhost" , 60000 ) ; final org . glassfish . grizzly . connectionpool . MultiEndpointPool < java . net . SocketAddress > pool = org . glassfish . grizzly . connectionpool . MultiEndpointPool . builder ( java . net . SocketAddress . class ) . maxConnectionsPerEndpoint ( 3 ) . maxConnectionsTotal ( 15 ) . keepAliveTimeout ( ( - 1 ) , TimeUnit . SECONDS ) . build ( ) ; final org . glassfish . grizzly . connectionpool . Endpoint < java . net . SocketAddress > key1 = Endpoint . Factory . create ( new java . net . InetSocketAddress ( "localhost" , org . glassfish . grizzly . connectionpool . MultiEndPointPoolTest . PORT ) , localAddress , transport ) ; try { org . glassfish . grizzly . Connection c1 = pool . take ( key1 ) . get ( ) ; "<AssertPlaceHolder>" ; } finally { pool . close ( ) ; } } getLocalAddress ( ) { return localSocketAddressHolder . get ( ) ; }
org . junit . Assert . assertEquals ( localAddress , c1 . getLocalAddress ( ) )
metadata_xmlComplexPropertyNullablefalse ( ) { final java . lang . String entryName = "bar/90_contents/odatacol1/00_$metadata.xml" ; final java . lang . String filename = "/00_$metadata_compprop_nullable_empty.xml" ; java . net . URL fileUrl = java . lang . ClassLoader . getSystemResource ( ( ( com . fujitsu . dc . test . unit . core . bar . BarFileValidateTest . RESOURCE_PATH ) + filename ) ) ; java . io . File file = new java . io . File ( fileUrl . getPath ( ) ) ; java . io . FileInputStream fis = null ; try { fis = new java . io . FileInputStream ( file ) ; com . fujitsu . dc . test . unit . core . bar . BarFileValidateTest . TestBarRunner testBarRunner = new com . fujitsu . dc . test . unit . core . bar . BarFileValidateTest . TestBarRunner ( ) ; boolean res = testBarRunner . registUserSchema ( entryName , fis , null ) ; "<AssertPlaceHolder>" ; return ; } catch ( com . fujitsu . dc . core . DcCoreException dce ) { org . junit . Assert . fail ( "Unexpected<sp>exception" ) ; } catch ( java . lang . Exception ex ) { org . junit . Assert . fail ( "Unexpected<sp>exception" ) ; } org . junit . Assert . fail ( "DcCoreException" ) ; } registUserSchema ( java . lang . String , java . io . InputStream , com . fujitsu . dc . core . model . DavCmp ) { org . odata4j . edm . EdmDataServices metadata = null ; try { java . io . InputStreamReader isr = new java . io . InputStreamReader ( new org . apache . commons . io . input . CloseShieldInputStream ( inputStream ) ) ; org . odata4j . stax2 . XMLFactoryProvider2 provider = org . odata4j . stax2 . staximpl . StaxXMLFactoryProvider2 . getInstance ( ) ; org . odata4j . stax2 . XMLInputFactory2 factory = provider . newXMLInputFactory2 ( ) ; org . odata4j . stax2 . XMLEventReader2 reader = factory . createXMLEventReader ( isr ) ; com . fujitsu . dc . core . odata . DcEdmxFormatParser parser = new com . fujitsu . dc . core . odata . DcEdmxFormatParser ( ) ; metadata = parser . parseMetadata ( reader ) ; } catch ( java . lang . Exception ex ) { com . fujitsu . dc . core . bar . BarFileReadRunner . log . info ( ( "XMLParseException:<sp>" + ( ex . getMessage ( ) ) ) , ex . fillInStackTrace ( ) ) ; java . lang . String message = com . fujitsu . dc . core . DcCoreMessageUtils . getMessage ( "PL-BI-2002" ) ; writeOutputStream ( true , "PL-BI-1004" , entryName , message ) ; return false ; } catch ( java . lang . StackOverflowError tw ) { com . fujitsu . dc . core . bar . BarFileReadRunner . log . info ( ( "XMLParseException:<sp>" + ( tw . getMessage ( ) ) ) , tw . fillInStackTrace ( ) ) ; java . lang . String message = com . fujitsu . dc . core . DcCoreMessageUtils . getMessage ( "PL-BI-2002" ) ; writeOutputStream ( true , "PL-BI-1004" , entryName , message ) ; return false ; } try { createComplexTypes ( metadata , davCmp ) ; createEntityTypes ( metadata , davCmp ) ; createAssociations ( metadata , davCmp ) ; } catch ( com . fujitsu . dc . core . DcCoreException e ) { writeOutputStream ( true , "PL-BI-1004" , entryName , e . getMessage ( ) ) ; com . fujitsu . dc . core . bar . BarFileReadRunner . log . info ( ( "DcCoreException:<sp>" + ( e . getMessage ( ) ) ) ) ; return false ; } catch ( java . lang . Exception e ) { com . fujitsu . dc . core . bar . BarFileReadRunner . log . info ( ( "Regist<sp>Entity<sp>Error:<sp>" + ( e . getMessage ( ) ) ) , e . fillInStackTrace ( ) ) ; java . lang . String message = com . fujitsu . dc . core . DcCoreMessageUtils . getMessage ( "PL-BI-2003" ) ; writeOutputStream ( true , "PL-BI-1004" , entryName , message ) ; return false ; } return true ; }
org . junit . Assert . assertFalse ( res )
testListDataRecordPaginatedStartingAtMinusTwo ( ) { com . liferay . data . engine . model . DEDataRecordCollection deDataRecordCollection = com . liferay . data . engine . service . test . DEDataEngineTestUtil . insertDEDataRecordCollection ( _adminUser , _group , _deDataDefinitionService , _deDataRecordCollectionService ) ; int total = 5 ; for ( int i = 0 ; i < total ; i ++ ) { com . liferay . data . engine . service . test . DEDataEngineTestUtil . insertDEDataRecord ( _adminUser , _group , deDataRecordCollection , _deDataRecordCollectionService ) ; } java . util . List < com . liferay . data . engine . model . DEDataRecord > deDataRecords = listDEDataRecords ( deDataRecordCollection . getDEDataRecordCollectionId ( ) , ( - 1 ) , 5 ) ; "<AssertPlaceHolder>" ; } toString ( ) { com . liferay . petra . string . StringBundler sb = new com . liferay . petra . string . StringBundler ( 23 ) ; sb . append ( ",<sp>width=" 1 ) ; sb . append ( uuid ) ; sb . append ( ",<sp>width=" 0 ) ; sb . append ( amImageEntryId ) ; sb . append ( ",<sp>groupId=" ) ; sb . append ( groupId ) ; sb . append ( ",<sp>companyId=" ) ; sb . append ( companyId ) ; sb . append ( ",<sp>createDate=" ) ; sb . append ( createDate ) ; sb . append ( ",<sp>configurationUuid=" ) ; sb . append ( configurationUuid ) ; sb . append ( ",<sp>fileVersionId=" ) ; sb . append ( fileVersionId ) ; sb . append ( ",<sp>mimeType=" ) ; sb . append ( mimeType ) ; sb . append ( ",<sp>height=" ) ; sb . append ( height ) ; sb . append ( ",<sp>width=" ) ; sb . append ( width ) ; sb . append ( ",<sp>size=" ) ; sb . append ( size ) ; sb . append ( "}" ) ; return sb . toString ( ) ; }
org . junit . Assert . assertEquals ( deDataRecords . toString ( ) , 5 , deDataRecords . size ( ) )
testGetV8ResultNull ( ) { java . lang . Object result = com . eclipsesource . v8 . utils . V8ObjectUtils . getV8Result ( v8 , null ) ; "<AssertPlaceHolder>" ; } getV8Result ( com . eclipsesource . v8 . V8 , java . lang . Object ) { if ( value == null ) { return null ; } java . util . Map < java . lang . Object , com . eclipsesource . v8 . V8Value > cache = new java . util . Hashtable < java . lang . Object , com . eclipsesource . v8 . V8Value > ( ) ; try { java . lang . Object result = com . eclipsesource . v8 . utils . V8ObjectUtils . getV8Result ( v8 , value , cache ) ; if ( result instanceof com . eclipsesource . v8 . V8Value ) { return ( ( com . eclipsesource . v8 . V8Value ) ( result ) ) . twin ( ) ; } return result ; } finally { for ( com . eclipsesource . v8 . V8Value v8Object : cache . values ( ) ) { v8Object . close ( ) ; } } }
org . junit . Assert . assertNull ( result )
testProcessData ( ) { java . lang . Object expectedObject = new java . lang . Object ( ) ; org . talend . codegen . flowvariables . runtime . MainDataProcessor mainDataProcessor = new org . talend . codegen . flowvariables . runtime . MainDataProcessor ( ) ; java . lang . Object actualObject = mainDataProcessor . processData ( expectedObject ) ; "<AssertPlaceHolder>" ; } processData ( java . lang . Object ) { return data ; }
org . junit . Assert . assertEquals ( expectedObject , actualObject )
testCreateFamilyMedicalHistorySectionBuilder ( ) { org . openhealthtools . mdht . uml . cda . builder . SectionBuilder < org . openhealthtools . mdht . uml . cda . ihe . FamilyMedicalHistorySection > sectionBuilder = org . openhealthtools . mdht . uml . cda . ihe . builder . IHEBuilderFactory . createFamilyMedicalHistorySectionBuilder ( ) ; org . openhealthtools . mdht . uml . cda . ihe . FamilyMedicalHistorySection section = sectionBuilder . buildSection ( ) ; "<AssertPlaceHolder>" ; Diagnostician . INSTANCE . validate ( section ) ; org . openhealthtools . mdht . uml . cda . util . CDAUtil . saveSnippet ( section , System . out ) ; } buildSection ( ) { org . openhealthtools . mdht . uml . cda . Section section = CDAFactory . eINSTANCE . createSection ( ) ; construct ( section ) ; return section ; }
org . junit . Assert . assertNotNull ( section )
testImplementingRepositoryInterfaceThroughOtherInterface ( ) { org . eclipse . jdt . core . IType type = javaProject . findType ( "org.ClassImplementingRepositoryInterfaceThroughOtherInterface" ) ; "<AssertPlaceHolder>" ; } isSpringDataRepository ( org . eclipse . jdt . core . IType ) { if ( type == null ) { return false ; } try { org . eclipse . jdt . core . IAnnotation [ ] annotations = type . getAnnotations ( ) ; for ( org . eclipse . jdt . core . IAnnotation annotation : annotations ) { if ( ( annotation . getElementName ( ) . equals ( "org.springframework.data.repository.NoRepositoryBean" ) ) || ( annotation . getElementName ( ) . equals ( "NoRepositoryBean" ) ) ) { return false ; } } return org . springframework . ide . eclipse . data . jdt . core . RepositoryInformation . isSpringDataRepositoryInterfaces ( type ) ; } catch ( org . eclipse . jdt . core . JavaModelException e ) { } return false ; }
org . junit . Assert . assertTrue ( org . springframework . ide . eclipse . data . jdt . core . RepositoryInformation . isSpringDataRepository ( type ) )
testCreateCacheIdAbsoluteURLWithPort ( ) { java . lang . String cacheId = com . hannikkala . poc . util . CacheIdUtil . createCacheId ( "http://localhost:3000/" , "http://localhost:3000/font-awesome@4.3.0/css/font-awesome.min.css" , "http://localhost:3000" ) ; "<AssertPlaceHolder>" ; } createCacheId ( java . lang . String , java . lang . String , java . lang . String ) { if ( ( ! ( baseUrl . endsWith ( "/" ) ) ) && ( ! ( file . startsWith ( "/" ) ) ) ) { baseUrl += "/" ; } java . net . URL url = null ; try { url = new java . net . URL ( new java . net . URL ( baseUrl ) , file ) ; if ( ( file . startsWith ( "http" ) ) && ( contextRoot . startsWith ( "http" ) ) ) { java . net . URL contextUrl = new java . net . URL ( contextRoot ) ; contextRoot = ( ( ( contextUrl . getProtocol ( ) ) + "://" ) + ( contextUrl . getHost ( ) ) ) + ( ( contextUrl . getPort ( ) ) != ( - 1 ) ? ":" + ( contextUrl . getPort ( ) ) : "" ) ; } } catch ( java . net . MalformedURLException e ) { com . hannikkala . poc . util . CacheIdUtil . _log . error ( ( ( ( "URL:<sp>" + baseUrl ) + file ) + "<sp>is<sp>malformed." ) , e ) ; } return contextRoot + ( url . getPath ( ) ) ; }
org . junit . Assert . assertEquals ( "http://localhost:3000/font-awesome@4.3.0/css/font-awesome.min.css" , cacheId )
findBetweenTwoDates ( ) { java . util . Calendar d = java . util . Calendar . getInstance ( ) ; d . add ( Calendar . DATE , ( - 1 ) ) ; topDomain = domainRepository . findById ( LoadingServiceTestDatas . sqlDomain ) ; rootDomain = domainRepository . findById ( LoadingServiceTestDatas . sqlRootDomain ) ; advancedStatisticMongoRepository . insert ( newMimeTypeStatistic ( topDomain , d . getTime ( ) ) ) ; advancedStatisticMongoRepository . insert ( newMimeTypeStatistic ( rootDomain , d . getTime ( ) ) ) ; java . util . Calendar date = java . util . Calendar . getInstance ( ) ; date . add ( Calendar . DATE , ( - 2 ) ) ; advancedStatisticMongoRepository . insert ( newMimeTypeStatistic ( topDomain , date . getTime ( ) ) ) ; java . util . List < org . linagora . linshare . core . domain . constants . ExceptionType > exceptionTypes = com . google . common . collect . Lists . newArrayList ( ) ; exceptionTypes . add ( ExceptionType . BUSINESS_EXCEPTION ) ; java . util . Calendar beginDate = java . util . Calendar . getInstance ( ) ; beginDate . add ( Calendar . DATE , ( - 3 ) ) ; java . util . Set < org . linagora . linshare . mongo . entities . MimeTypeStatistic > mimeTypeStatistics = advancedStatisticMongoRepository . findBetweenTwoDates ( topDomain . getUuid ( ) , beginDate . getTime ( ) , new java . util . Date ( ) , "application/pdf" ) ; "<AssertPlaceHolder>" ; } getUuid ( ) { return uuid ; }
org . junit . Assert . assertEquals ( 2 , mimeTypeStatistics . size ( ) )
returnsInstanceGivenInCreationOfObjectFactory ( ) { java . lang . Object instance = new java . lang . Object ( ) ; org . everrest . core . ObjectModel objectModel = mock ( org . everrest . core . ObjectModel . class ) ; org . everrest . core . SingletonObjectFactory singletonObjectFactory = new org . everrest . core . SingletonObjectFactory ( objectModel , instance ) ; java . lang . Object result = singletonObjectFactory . getInstance ( null ) ; "<AssertPlaceHolder>" ; } getInstance ( org . everrest . core . ApplicationContext ) { try { java . lang . Object object = resourceField . get ( testParent ) ; if ( object != null ) { org . everrest . core . provider . ProviderDescriptor descriptor = new org . everrest . core . impl . provider . ProviderDescriptorImpl ( object ) ; java . util . List < org . everrest . core . FieldInjector > fieldInjectors = model . getFieldInjectors ( ) ; if ( ( fieldInjectors != null ) && ( ( fieldInjectors . size ( ) ) > 0 ) ) { fieldInjectors . stream ( ) . filter ( ( injector ) -> ( injector . getAnnotation ( ) ) != null ) . forEach ( ( injector ) -> injector . inject ( object , context ) ) ; } return new org . everrest . core . SingletonObjectFactory ( descriptor , object ) . getInstance ( context ) ; } else { org . everrest . core . provider . ProviderDescriptor descriptor = new org . everrest . core . impl . provider . ProviderDescriptorImpl ( resourceField . getType ( ) ) ; return new org . everrest . core . PerRequestObjectFactory ( descriptor ) . getInstance ( context ) ; } } catch ( java . lang . IllegalArgumentException e ) { org . everrest . assured . TestResourceFactory . LOG . error ( e . getLocalizedMessage ( ) , e ) ; throw new java . lang . RuntimeException ( e . getLocalizedMessage ( ) , e ) ; } catch ( java . lang . IllegalAccessException e ) { org . everrest . assured . TestResourceFactory . LOG . error ( e . getLocalizedMessage ( ) , e ) ; throw new java . lang . RuntimeException ( e . getLocalizedMessage ( ) , e ) ; } }
org . junit . Assert . assertSame ( instance , result )
testGetTreePermission ( ) { org . apache . jackrabbit . oak . spi . security . authorization . permission . PermissionProvider pp = createPermissionProvider ( ) ; org . apache . jackrabbit . oak . api . Tree t = root . getTree ( PermissionConstants . PERMISSIONS_STORE_PATH ) ; "<AssertPlaceHolder>" ; } getTreePermission ( org . apache . jackrabbit . oak . api . Tree , org . apache . jackrabbit . oak . spi . security . authorization . permission . TreePermission ) { return TreePermission . EMPTY ; }
org . junit . Assert . assertSame ( TreePermission . EMPTY , pp . getTreePermission ( t , TreePermission . EMPTY ) )
testUpdate ( ) { org . smartdata . model . ClusterInfo clusterInfo = new org . smartdata . model . ClusterInfo ( ) ; clusterInfo . setCid ( 1 ) ; clusterInfo . setType ( "test" ) ; clusterInfo . setState ( "test" ) ; clusterInfo . setConfPath ( "test" ) ; clusterInfo . setUrl ( "test" ) ; clusterInfo . setName ( "test" ) ; clusterInfoDao . insert ( clusterInfo ) ; clusterInfo . setState ( "test1" ) ; clusterInfo . setType ( "test1" ) ; clusterInfoDao . updateState ( 1 , "test1" ) ; clusterInfoDao . updateType ( 1 , "test1" ) ; "<AssertPlaceHolder>" ; } getById ( long ) { org . springframework . jdbc . core . JdbcTemplate jdbcTemplate = new org . springframework . jdbc . core . JdbcTemplate ( dataSource ) ; return jdbcTemplate . queryForObject ( "SELECT<sp>*<sp>FROM<sp>file<sp>WHERE<sp>fid<sp>=<sp>?" , new java . lang . Object [ ] { fid } , new org . smartdata . metastore . dao . FileInfoDao . FileInfoRowMapper ( ) ) ; }
org . junit . Assert . assertTrue ( clusterInfoDao . getById ( 1 ) . equals ( clusterInfo ) )
compareDefaultSortOrderFileToString ( ) { java . lang . String expected = org . apache . commons . io . IOUtils . toString ( new java . io . FileInputStream ( "src/test/resources/sortOrderFiles/with_newline_tagsToString.txt" ) , sortpom . wrapper . CustomSortOrderFileTest . UTF_8 ) ; "<AssertPlaceHolder>" ; } getToStringOnCustomSortOrderFile ( ) { sortpom . parameter . PluginParameters pluginParameters = sortpom . parameter . PluginParameters . builder ( ) . setPomFile ( null ) . setFileOutput ( false , ".bak" , null ) . setEncoding ( "UTF-8" ) . setFormatting ( "\r\n" , true , true ) . setIndent ( 2 , false ) . setSortOrder ( "src/test/resources/sortOrderFiles/with_newline_tags.xml" , null ) . setSortEntities ( "scope,groupId,artifactId" , "groupId,artifactId" , true , true ) . build ( ) ; sortpom . util . FileUtil fileUtil = new sortpom . util . FileUtil ( ) ; fileUtil . setup ( pluginParameters ) ; sortpom . wrapper . WrapperFactoryImpl wrapperFactory = new sortpom . wrapper . WrapperFactoryImpl ( fileUtil ) ; org . jdom . Document documentFromDefaultSortOrderFile = wrapperFactory . createDocumentFromDefaultSortOrderFile ( ) ; new sortpom . wrapper . operation . HierarchyRootWrapper ( wrapperFactory . create ( documentFromDefaultSortOrderFile . getRootElement ( ) ) ) ; sortpom . wrapper . operation . HierarchyRootWrapper rootWrapper = new sortpom . wrapper . operation . HierarchyRootWrapper ( wrapperFactory . create ( documentFromDefaultSortOrderFile . getRootElement ( ) ) ) ; rootWrapper . createWrappedStructure ( wrapperFactory ) ; return rootWrapper . toString ( ) ; }
org . junit . Assert . assertEquals ( expected , getToStringOnCustomSortOrderFile ( ) )
isValidProfileHashNullRequestNullTest ( ) { org . kaaproject . kaa . server . sync . ClientSyncMetaData clientSyncMetaData = mock ( org . kaaproject . kaa . server . sync . ClientSyncMetaData . class ) ; when ( clientSyncMetaData . getProfileHash ( ) ) . thenReturn ( null ) ; org . kaaproject . kaa . server . sync . ClientSync clientSync = new org . kaaproject . kaa . server . sync . ClientSync ( 0 , clientSyncMetaData , null , null , null , null , null , null ) ; "<AssertPlaceHolder>" ; } isValid ( ) { org . kaaproject . kaa . server . sync . ClientSyncMetaData md = this . getClientSyncMetaData ( ) ; if ( ( md . getProfileHash ( ) ) == null ) { org . kaaproject . kaa . server . sync . ProfileClientSync profileRequest = this . getProfileSync ( ) ; if ( ( profileRequest == null ) || ( ( profileRequest . getEndpointPublicKey ( ) ) == null ) ) { return false ; } } return true ; }
org . junit . Assert . assertEquals ( clientSync . isValid ( ) , false )
testCurrentMappingXML ( ) { com . github . dozermapper . core . builder . xml . JAXBModelParser < com . github . dozermapper . core . builder . model . jaxb . MappingsDefinition > factory = new com . github . dozermapper . core . builder . xml . DefaultJAXBModelParser ( new com . github . dozermapper . core . config . BeanContainer ( ) ) ; java . io . File folder = new java . io . File ( ( ( new java . io . File ( "." ) . getCanonicalPath ( ) ) + "/src/test/resources/mappings" ) ) ; java . io . File [ ] listOfFiles = folder . listFiles ( ) ; com . github . dozermapper . core . builder . xml . DefaultJAXBModelParserTest . LOG . info ( "Found<sp>{}<sp>to<sp>validate." , listOfFiles . length ) ; for ( java . io . File file : listOfFiles ) { if ( file . isFile ( ) ) { com . github . dozermapper . core . builder . xml . DefaultJAXBModelParserTest . LOG . info ( "Validating<sp>{}" , file . getAbsoluteFile ( ) ) ; try ( java . io . FileInputStream fileInputStream = new java . io . FileInputStream ( file ) ) { java . lang . String xmlContent = org . apache . commons . io . IOUtils . toString ( fileInputStream , java . nio . charset . Charset . forName ( "UTF-8" ) ) ; com . github . dozermapper . core . builder . model . jaxb . MappingsDefinition mapping = factory . readXML ( xmlContent , com . github . dozermapper . core . builder . model . jaxb . MappingsDefinition . class ) ; "<AssertPlaceHolder>" ; factory . validateXML ( xmlContent ) ; } } } } readXML ( java . lang . String , java . lang . Class ) { javax . xml . bind . JAXBElement < T > answer ; try ( java . io . StringReader stringReader = new java . io . StringReader ( xmlContent ) ) { javax . xml . transform . stream . StreamSource streamSource = new javax . xml . transform . stream . StreamSource ( stringReader ) ; javax . xml . bind . JAXBContext jaxbContext = getOrCreateJAXBContext ( ) ; javax . xml . bind . Unmarshaller jaxbUnmarshaller = jaxbContext . createUnmarshaller ( ) ; answer = jaxbUnmarshaller . unmarshal ( streamSource , type ) ; } catch ( javax . xml . bind . JAXBException ex ) { throw new org . xml . sax . SAXException ( ex ) ; } return answer . getValue ( ) ; }
org . junit . Assert . assertNotNull ( mapping )
testIsRightArrowWhenItIsRightArrow ( ) { final elemental2 . dom . Element element = mock ( elemental2 . dom . Element . class ) ; final elemental2 . dom . DOMTokenList classList = mock ( elemental2 . dom . DOMTokenList . class ) ; element . classList = classList ; when ( classList . contains ( org . kie . workbench . common . dmn . client . editors . types . listview . common . ListItemViewCssHelper . RIGHT_ARROW_CSS_CLASS ) ) . thenReturn ( true ) ; "<AssertPlaceHolder>" ; } isRightArrow ( elemental2 . dom . Element ) { return element . classList . contains ( org . kie . workbench . common . dmn . client . editors . types . listview . common . ListItemViewCssHelper . RIGHT_ARROW_CSS_CLASS ) ; }
org . junit . Assert . assertTrue ( org . kie . workbench . common . dmn . client . editors . types . listview . common . ListItemViewCssHelper . isRightArrow ( element ) )
testConvertToKettleValueRecordStringValue ( ) { avroField . m_kettleType = "String" ; avroField . m_fieldPath = "key" ; avroField . init ( 0 ) ; avroField . reset ( variableSpace ) ; java . lang . String testString = "testString" ; org . apache . avro . Schema schema = mock ( org . apache . avro . Schema . class ) ; org . apache . avro . Schema . Field field = mock ( Schema . Field . class ) ; when ( schema . getFields ( ) ) . thenReturn ( new java . util . ArrayList ( java . util . Arrays . asList ( field ) ) ) ; when ( schema . getField ( avroField . m_fieldPath ) ) . thenReturn ( field ) ; when ( schema . getType ( ) ) . thenReturn ( Schema . Type . RECORD ) ; org . apache . avro . Schema fieldSchema = mock ( org . apache . avro . Schema . class ) ; when ( field . schema ( ) ) . thenReturn ( fieldSchema ) ; when ( fieldSchema . getType ( ) ) . thenReturn ( Schema . Type . STRING ) ; org . apache . avro . generic . GenericData . Record record = new org . apache . avro . generic . GenericData . Record ( schema ) ; record . put ( avroField . m_fieldPath , testString ) ; "<AssertPlaceHolder>" ; } convertToKettleValue ( org . apache . avro . generic . GenericData$Record , org . apache . avro . Schema , org . apache . avro . Schema , boolean ) { if ( record == null ) { return null ; } if ( ( m_tempParts . size ( ) ) == 0 ) { throw new org . pentaho . di . core . exception . KettleException ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.MalformedPathRecord" ) ) ; } java . lang . String part = m_tempParts . remove ( 0 ) ; if ( ( part . charAt ( 0 ) ) == '[' ) { throw new org . pentaho . di . core . exception . KettleException ( ( ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.InvalidPath" ) ) + ( m_tempParts ) ) ) ; } if ( ( part . indexOf ( '[' ) ) > 0 ) { java . lang . String arrayPart = part . substring ( part . indexOf ( '[' ) ) ; part = part . substring ( 0 , part . indexOf ( '[' ) ) ; m_tempParts . add ( 0 , arrayPart ) ; } org . apache . avro . Schema . Field fieldS = s . getField ( part ) ; if ( ( fieldS == null ) && ( ! ignoreMissing ) ) { throw new org . pentaho . di . core . exception . KettleException ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.NonExistentField" , part ) ) ; } java . lang . Object field = record . get ( part ) ; if ( field == null ) { fieldS = defaultSchema . getField ( part ) ; if ( ( fieldS == null ) || ( ( fieldS . defaultValue ( ) ) == null ) ) { return null ; } field = fieldS . defaultValue ( ) ; } org . apache . avro . Schema . Type fieldT = fieldS . schema ( ) . getType ( ) ; org . apache . avro . Schema fieldSchema = fieldS . schema ( ) ; if ( fieldT == ( Schema . Type . UNION ) ) { if ( field instanceof org . apache . avro . generic . GenericContainer ) { fieldSchema = ( ( org . apache . avro . generic . GenericContainer ) ( field ) ) . getSchema ( ) ; fieldT = fieldSchema . getType ( ) ; } else { if ( field instanceof java . util . Map ) { org . apache . avro . Schema mapSchema = null ; for ( org . apache . avro . Schema ts : fieldSchema . getTypes ( ) ) { if ( ( ts . getType ( ) ) == ( Schema . Type . MAP ) ) { mapSchema = ts ; break ; } } if ( mapSchema == null ) { throw new org . pentaho . di . core . exception . KettleException ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.UnableToFindSchemaForUnionMap" ) ) ; } fieldSchema = mapSchema ; fieldT = Schema . Type . MAP ; } else { if ( ( m_tempValueMeta . getType ( ) ) != ( org . pentaho . di . core . row . ValueMetaInterface . TYPE_STRING ) ) { fieldSchema = org . pentaho . di . trans . steps . avroinput . AvroInputData . checkUnion ( fieldSchema ) ; fieldT = fieldSchema . getType ( ) ; } else { fieldSchema = org . apache . avro . Schema . create ( Schema . Type . STRING ) ; fieldT = fieldSchema . getType ( ) ; } } } } if ( fieldT == ( Schema . Type . RECORD ) ) { return convertToKettleValue ( ( ( org . apache . avro . generic . GenericData . Record ) ( field ) ) , fieldSchema , defaultSchema , ignoreMissing ) ; }
org . junit . Assert . assertEquals ( testString , avroField . convertToKettleValue ( record , schema , mock ( org . apache . avro . Schema . class ) , false ) )
testResolveLocale ( ) { org . apache . tiles . request . Request request = createMock ( org . apache . tiles . request . Request . class ) ; java . util . Map < java . lang . String , java . lang . Object > sessionScope = createMock ( java . util . Map . class ) ; java . util . Locale locale = java . util . Locale . ITALY ; expect ( request . getContext ( "session" ) ) . andReturn ( sessionScope ) ; expect ( sessionScope . get ( DefaultLocaleResolver . LOCALE_KEY ) ) . andReturn ( null ) ; expect ( request . getRequestLocale ( ) ) . andReturn ( locale ) ; replay ( request , sessionScope ) ; org . apache . tiles . locale . impl . DefaultLocaleResolver resolver = new org . apache . tiles . locale . impl . DefaultLocaleResolver ( ) ; "<AssertPlaceHolder>" ; verify ( request , sessionScope ) ; } resolveLocale ( org . apache . tiles . request . Request ) { java . util . Locale retValue = null ; java . util . Map < java . lang . String , java . lang . Object > session = request . getContext ( "session" ) ; if ( session != null ) { retValue = ( ( java . util . Locale ) ( session . get ( org . apache . tiles . locale . impl . DefaultLocaleResolver . LOCALE_KEY ) ) ) ; } if ( retValue == null ) { retValue = request . getRequestLocale ( ) ; } return retValue ; }
org . junit . Assert . assertSame ( locale , resolver . resolveLocale ( request ) )
testReadPoints ( ) { java . lang . String namedCurve = "secp192r1" ; java . util . List < de . rub . nds . tlsattacker . attacks . ec . ICEPoint > result = de . rub . nds . tlsattacker . attacks . ec . ICEPointReader . readPoints ( namedCurve ) ; "<AssertPlaceHolder>" ; } getOrder ( ) { return order ; }
org . junit . Assert . assertEquals ( 5 , result . get ( 0 ) . getOrder ( ) )
testCoerceToType02 ( ) { java . lang . Object result = javax . el . ELManager . getExpressionFactory ( ) . coerceToType ( null , int . class ) ; "<AssertPlaceHolder>" ; } coerceToType ( java . lang . Class , java . lang . String ) { if ( type . equals ( java . lang . String . class ) ) { return value ; } else if ( ( type . equals ( boolean . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Boolean . class ) ) ) { return java . lang . Boolean . valueOf ( value ) ; } else if ( ( type . equals ( byte . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Byte . class ) ) ) { return java . lang . Byte . valueOf ( value ) ; } else if ( ( ( value . length ( ) ) == 1 ) && ( ( type . equals ( char . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Character . class ) ) ) ) { return java . lang . Character . valueOf ( value . charAt ( 0 ) ) ; } else if ( ( type . equals ( double . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Double . class ) ) ) { return java . lang . Double . valueOf ( value ) ; } else if ( ( type . equals ( float . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Float . class ) ) ) { return java . lang . Float . valueOf ( value ) ; } else if ( ( type . equals ( int . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Integer . class ) ) ) { return java . lang . Integer . valueOf ( value ) ; } else if ( ( type . equals ( long . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Long . class ) ) ) { return java . lang . Long . valueOf ( value ) ; } else if ( ( type . equals ( short . class ) ) || ( type . equals ( org . apache . tomcat . websocket . Short . class ) ) ) { return java . lang . Short . valueOf ( value ) ; } else { throw new java . lang . IllegalArgumentException ( org . apache . tomcat . websocket . Util . sm . getString ( "util.invalidType" , value , type . getName ( ) ) ) ; } }
org . junit . Assert . assertEquals ( java . lang . Integer . valueOf ( 0 ) , result )
testSetCommand ( ) { org . opencastproject . dictionary . hunspell . DictionaryServiceImpl service = new org . opencastproject . dictionary . hunspell . DictionaryServiceImpl ( ) ; java . lang . String command = "123" ; service . setCommand ( command ) ; "<AssertPlaceHolder>" ; } getCommand ( ) { return command ; }
org . junit . Assert . assertEquals ( command , service . getCommand ( ) )
testJiraAwareFailureRegex ( ) { JiraAwareFailureUtils . regexJiraMatches . put ( "TEST.*UNIQUE.*TEST" , "KULRICE-TEST" ) ; org . kuali . rice . testtools . common . JiraAwareFailureUtils . failOnMatchedJira ( "TEST<sp>UNIQUE<sp>data<sp>for<sp>TEST" , this ) ; "<AssertPlaceHolder>" ; } contains ( java . lang . Class ) { return ( getMetadata ( type ) ) != null ; }
org . junit . Assert . assertTrue ( message . contains ( "KULRICE-TEST" ) )
testTokenGeneratorWithStandardUsername ( ) { axiom . delauth . token . TokenGenerator generator = new axiom . delauth . token . TokenGenerator ( ) ; final java . lang . String USERNAME = "testUsername" ; final java . lang . String TOKEN = generator . generateToken ( USERNAME ) ; "<AssertPlaceHolder>" ; } generateToken ( java . lang . String ) { java . lang . String token = java . lang . Integer . toHexString ( new java . util . Random ( ) . nextInt ( Integer . MAX_VALUE ) ) ; tokenStore . addToken ( username , token ) ; return token ; }
org . junit . Assert . assertNotNull ( TOKEN )
testDoDestroy ( ) { control . bind ( session ) ; control . doDestroy ( ) ; "<AssertPlaceHolder>" ; } getGridPanel ( ) { return ( ( org . kie . workbench . common . dmn . client . widgets . panel . DMNGridPanelControl ) ( getSession ( ) . getCanvasControl ( org . kie . workbench . common . dmn . client . widgets . panel . DMNGridPanelControl . class ) ) ) . getGridPanel ( ) ; }
org . junit . Assert . assertNull ( control . getGridPanel ( ) )