input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testEncryptDecryptAesGcmWrapA128CBCHS256 ( ) { if ( "IBM<sp>Corporation" . equals ( java . lang . System . getProperty ( "java.vendor" ) ) ) { return ; } final java . lang . String specPlainText = "Live<sp>long<sp>and<sp>prosper." ; byte [ ] cekEncryptionKey = org . apache . cxf . common . util . Base64UrlUtility . decode ( org . apache . cxf . rs . security . jose . jwe . JweCompactReaderWriterTest . KEY_ENCRYPTION_KEY_A3 ) ; org . apache . cxf . rs . security . jose . jwe . AesGcmWrapKeyEncryptionAlgorithm keyEncryption = new org . apache . cxf . rs . security . jose . jwe . AesGcmWrapKeyEncryptionAlgorithm ( cekEncryptionKey , org . apache . cxf . rs . security . jose . jwa . KeyAlgorithm . A128GCMKW ) ; org . apache . cxf . rs . security . jose . jwe . JweEncryptionProvider encryption = new org . apache . cxf . rs . security . jose . jwe . AesCbcHmacJweEncryption ( org . apache . cxf . rs . security . jose . jwa . ContentAlgorithm . A128CBC_HS256 , org . apache . cxf . rs . security . jose . jwe . JweCompactReaderWriterTest . CONTENT_ENCRYPTION_KEY_A3 , org . apache . cxf . rs . security . jose . jwe . JweCompactReaderWriterTest . INIT_VECTOR_A3 , keyEncryption ) ; java . lang . String jweContent = encryption . encrypt ( specPlainText . getBytes ( StandardCharsets . UTF_8 ) , null ) ; org . apache . cxf . rs . security . jose . jwe . AesGcmWrapKeyDecryptionAlgorithm keyDecryption = new org . apache . cxf . rs . security . jose . jwe . AesGcmWrapKeyDecryptionAlgorithm ( cekEncryptionKey ) ; org . apache . cxf . rs . security . jose . jwe . JweDecryptionProvider decryption = new org . apache . cxf . rs . security . jose . jwe . AesCbcHmacJweDecryption ( keyDecryption ) ; java . lang . String decryptedText = decryption . decrypt ( jweContent ) . getContentText ( ) ; "<AssertPlaceHolder>" ; } getContentText ( ) { return new java . lang . String ( getContent ( ) , java . nio . charset . StandardCharsets . UTF_8 ) ; }
org . junit . Assert . assertEquals ( specPlainText , decryptedText )
getValueFromJsonObjectValueNull ( ) { java . lang . String jsonString = "{\"deviceId\":\"xxx-device\",\"generationId\":null,\"etag\":\"MA==\",\"connectionState\":\"Disconnected\",\"status\":\"Disabled\",\"statusReason\":null,\"connectionStateUpdatedTime\":\"0001-01-01T00:00:00\",\"statusUpdatedTime\":\"0001-01-01T00:00:00\",\"lastActivityTime\":\"0001-01-01T00:00:00\",\"cloudToDeviceMessageCount\":0,\"authentication\":{\"symmetricKey\":{\"primaryKey\":\"AAABBBCCC111222333444000\",\"secondaryKey\":\"111222333444555AAABBBCCC\"}}}" ; java . io . StringReader stringReader = new java . io . StringReader ( jsonString ) ; javax . json . JsonReader jsonReader = javax . json . Json . createReader ( stringReader ) ; javax . json . JsonObject jsonObject = jsonReader . readObject ( ) ; java . lang . String key = "generationId" ; java . lang . String expResult = "" ; java . lang . String result = com . microsoft . azure . sdk . iot . deps . util . Tools . getValueFromJsonObject ( jsonObject , key ) ; "<AssertPlaceHolder>" ; } getValueFromJsonObject ( javax . json . JsonObject , java . lang . String ) { java . lang . String retVal ; if ( ( ( ( jsonObject == null ) || ( jsonObject == ( javax . json . JsonObject . NULL ) ) ) || ( key == null ) ) || ( ( key . length ( ) ) == 0 ) ) { retVal = "" ; } else { javax . json . JsonValue jsonValue = jsonObject . get ( key ) ; if ( jsonValue != ( javax . json . JsonValue . NULL ) ) { retVal = com . microsoft . azure . sdk . iot . deps . util . Tools . getValueFromJsonString ( jsonObject . getJsonString ( key ) ) ; } else { retVal = "" ; } } return retVal ; }
org . junit . Assert . assertEquals ( expResult , result )
mGetShouldReturnMultipleSameKeysWhenKeysMapToSameSlot ( ) { nativeConnection . set ( org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . SAME_SLOT_KEY_1_BYTES , org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . VALUE_1_BYTES ) ; nativeConnection . set ( org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . SAME_SLOT_KEY_2_BYTES , org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . VALUE_2_BYTES ) ; org . springframework . data . redis . connection . jedis . List < byte [ ] > result = clusterConnection . mGet ( org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . SAME_SLOT_KEY_1_BYTES , org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . SAME_SLOT_KEY_2_BYTES , org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . SAME_SLOT_KEY_1_BYTES ) ; "<AssertPlaceHolder>" ; } set ( int , E ) { E object = get ( index ) ; listOps . set ( index , e ) ; return object ; }
org . junit . Assert . assertThat ( result , contains ( org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . VALUE_1_BYTES , org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . VALUE_2_BYTES , org . springframework . data . redis . connection . jedis . JedisClusterConnectionTests . VALUE_1_BYTES ) )
testVerschillendeDatums4 ( ) { java . util . List < nl . bzk . brp . model . validatie . Melding > meldingen = brby0527 . executeer ( null , maakPersoonMetPlaats ( maakGemeente ( ( ( short ) ( 24 ) ) , "gemeente" , 19920101 , 200020101 ) ) , maakActie ( "id.actie1" , 19910231 , 20070231 ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , meldingen . size ( ) )
testContentTypes ( ) { final com . sebastian_daschner . jaxrs_analyzer . analysis . results . Resources expectedResult = new com . sebastian_daschner . jaxrs_analyzer . analysis . results . Resources ( ) ; expectedResult . setBasePath ( "path" ) ; final com . sebastian_daschner . jaxrs_analyzer . analysis . results . ResourceMethod resourceGetMethod = com . sebastian_daschner . jaxrs_analyzer . analysis . results . ResourceMethodBuilder . withMethod ( HttpMethod . GET ) . andResponse ( 200 , com . sebastian_daschner . jaxrs_analyzer . analysis . results . ResponseBuilder . withResponseBody ( com . sebastian_daschner . jaxrs_analyzer . analysis . results . TypeUtils . STRING_IDENTIFIER ) . build ( ) ) . andResponseMediaTypes ( "application/xml" ) . build ( ) ; expectedResult . addMethod ( "test" , resourceGetMethod ) ; final com . sebastian_daschner . jaxrs_analyzer . model . results . ClassResult appPathResult = com . sebastian_daschner . jaxrs_analyzer . analysis . results . ClassResultBuilder . withApplicationPath ( "path" ) . build ( ) ; final com . sebastian_daschner . jaxrs_analyzer . model . results . MethodResult method = com . sebastian_daschner . jaxrs_analyzer . analysis . results . MethodResultBuilder . withResponses ( com . sebastian_daschner . jaxrs_analyzer . analysis . results . HttpResponseBuilder . withStatues ( 200 ) . andEntityTypes ( Types . STRING ) . build ( ) ) . andMethod ( HttpMethod . GET ) . andResponseMediaTypes ( "application/xml" ) . build ( ) ; final com . sebastian_daschner . jaxrs_analyzer . model . results . ClassResult resClassResult = com . sebastian_daschner . jaxrs_analyzer . analysis . results . ClassResultBuilder . withResourcePath ( "test" ) . andMethods ( method ) . build ( ) ; final java . util . Set < com . sebastian_daschner . jaxrs_analyzer . model . results . ClassResult > results = new java . util . HashSet ( java . util . Arrays . asList ( appPathResult , resClassResult ) ) ; final com . sebastian_daschner . jaxrs_analyzer . analysis . results . Resources actualResult = classUnderTest . interpret ( results ) ; "<AssertPlaceHolder>" ; } interpret ( java . util . Set ) { resources = new com . sebastian_daschner . jaxrs_analyzer . model . rest . Resources ( ) ; resources . setBasePath ( com . sebastian_daschner . jaxrs_analyzer . analysis . results . PathNormalizer . getApplicationPath ( classResults ) ) ; javaTypeAnalyzer = new com . sebastian_daschner . jaxrs_analyzer . analysis . results . JavaTypeAnalyzer ( resources . getTypeRepresentations ( ) ) ; dynamicTypeAnalyzer = new com . sebastian_daschner . jaxrs_analyzer . analysis . results . DynamicTypeAnalyzer ( resources . getTypeRepresentations ( ) ) ; stringParameterResolver = new com . sebastian_daschner . jaxrs_analyzer . analysis . results . StringParameterResolver ( resources . getTypeRepresentations ( ) , javaTypeAnalyzer ) ; classResults . stream ( ) . filter ( ( c ) -> ( c . getResourcePath ( ) ) != null ) . forEach ( this :: interpretClassResult ) ; resources . consolidateMultiplePaths ( ) ; return resources ; }
org . junit . Assert . assertEquals ( expectedResult , actualResult )
testChannelClosedOnExceptionDuringErrorNotification ( ) { org . apache . flink . shaded . netty4 . io . netty . channel . embedded . EmbeddedChannel ch = createEmbeddedChannel ( ) ; org . apache . flink . runtime . io . network . NetworkClientHandler handler = getClientHandler ( ch ) ; org . apache . flink . runtime . io . network . partition . consumer . RemoteInputChannel rich = addInputChannel ( handler ) ; doThrow ( new java . lang . RuntimeException ( "Expected<sp>test<sp>exception" ) ) . when ( rich ) . onError ( any ( java . lang . Throwable . class ) ) ; ch . pipeline ( ) . fireExceptionCaught ( new java . lang . Exception ( ) ) ; "<AssertPlaceHolder>" ; } isActive ( ) { return ! ( isIdle ( ) ) ; }
org . junit . Assert . assertFalse ( ch . isActive ( ) )
testPut ( ) { org . talend . components . jira . avro . IssueIndexedRecord indexedRecord = new org . talend . components . jira . avro . IssueIndexedRecord ( org . talend . components . jira . avro . IssueIndexedRecordTest . testJson , null ) ; indexedRecord . put ( 0 , org . talend . components . jira . avro . IssueIndexedRecordTest . testJson ) ; "<AssertPlaceHolder>" ; } get ( int ) { return values [ i ] ; }
org . junit . Assert . assertEquals ( org . talend . components . jira . avro . IssueIndexedRecordTest . testJson , indexedRecord . get ( 0 ) )
testGetFileName ( ) { System . out . println ( "getFileName" ) ; kg . apc . jmeter . config . VariablesFromCSV instance = new kg . apc . jmeter . config . VariablesFromCSV ( ) ; java . lang . String expResult = "" ; java . lang . String result = instance . getFileName ( ) ; "<AssertPlaceHolder>" ; } getFileName ( ) { return getPropertyAsString ( kg . apc . jmeter . config . VariablesFromCSV . FILENAME ) ; }
org . junit . Assert . assertEquals ( expResult , result )
testToIntArray ( ) { intSet . add ( 1 ) ; intSet . add ( 4 ) ; int [ ] array = intSet . toIntArray ( ) ; "<AssertPlaceHolder>" ; } toIntArray ( ) { int [ ] array = new int [ 1 ] ; array [ 0 ] = value ; return array ; }
org . junit . Assert . assertArrayEquals ( new int [ ] { 1 , 4 } , array )
shouldConvertMapToClassKeysWithProvidedMap ( ) { final java . util . Map < java . lang . String , java . lang . String > map = new java . util . HashMap ( ) ; map . put ( uk . gov . gchq . gaffer . commonutil . Integer . class . getName ( ) , "integer" ) ; map . put ( uk . gov . gchq . gaffer . commonutil . Double . class . getName ( ) , "double" ) ; map . put ( uk . gov . gchq . gaffer . commonutil . Long . class . getName ( ) , "long" ) ; final java . util . Map < java . lang . Class < ? extends java . lang . Number > , java . lang . String > result = new java . util . LinkedHashMap ( ) ; uk . gov . gchq . gaffer . commonutil . CollectionUtil . toMapWithClassKeys ( map , result ) ; final java . util . Map < java . lang . Class < ? extends java . lang . Number > , java . lang . String > expectedResult = new java . util . LinkedHashMap ( ) ; expectedResult . put ( uk . gov . gchq . gaffer . commonutil . Integer . class , "integer" ) ; expectedResult . put ( uk . gov . gchq . gaffer . commonutil . Double . class , "double" ) ; expectedResult . put ( uk . gov . gchq . gaffer . commonutil . Long . class , "long" ) ; "<AssertPlaceHolder>" ; } put ( uk . gov . gchq . gaffer . graph . GraphSerialisable , uk . gov . gchq . gaffer . federatedstore . FederatedAccess ) { if ( graph != null ) { java . lang . String graphId = graph . getDeserialisedConfig ( ) . getGraphId ( ) ; try { if ( null == access ) { throw new java . lang . IllegalArgumentException ( uk . gov . gchq . gaffer . federatedstore . FederatedGraphStorage . ACCESS_IS_NULL ) ; } if ( null != ( graphLibrary ) ) { graphLibrary . checkExisting ( graphId , graph . getDeserialisedSchema ( ) , graph . getDeserialisedProperties ( ) ) ; } validateExisting ( graph ) ; final uk . gov . gchq . gaffer . graph . Graph builtGraph = graph . getGraph ( ) ; if ( isCacheEnabled ( ) ) { addToCache ( builtGraph , access ) ; } java . util . Set < uk . gov . gchq . gaffer . graph . Graph > existingGraphs = storage . get ( access ) ; if ( null == existingGraphs ) { existingGraphs = com . google . common . collect . Sets . newHashSet ( builtGraph ) ; storage . put ( access , existingGraphs ) ; } else { existingGraphs . add ( builtGraph ) ; } } catch ( final java . lang . Exception e ) { throw new uk . gov . gchq . gaffer . federatedstore . exception . StorageException ( ( ( ( "Error<sp>adding<sp>graph<sp>" + graphId ) + "<sp>to<sp>storage<sp>due<sp>to:<sp>" ) + ( e . getMessage ( ) ) ) , e ) ; } } else { throw new uk . gov . gchq . gaffer . federatedstore . exception . StorageException ( "Graph<sp>cannot<sp>be<sp>null" ) ; } }
org . junit . Assert . assertEquals ( expectedResult , result )
testSetParent ( ) { org . libreplan . business . calendars . entities . BaseCalendar calendar = org . libreplan . business . test . calendars . entities . BaseCalendarTest . createBasicCalendar ( ) ; org . libreplan . business . calendars . entities . BaseCalendar calendar2 = org . libreplan . business . test . calendars . entities . BaseCalendarTest . createBasicCalendar ( ) ; org . libreplan . business . calendars . entities . BaseCalendar derived = calendar . newDerivedCalendar ( ) ; derived . setParent ( calendar2 ) ; "<AssertPlaceHolder>" ; } getParent ( ) { return parent ; }
org . junit . Assert . assertThat ( derived . getParent ( ) , org . hamcrest . CoreMatchers . equalTo ( calendar2 ) )
testReadNdefMessagesEmpty ( ) { for ( org . nfctools . mf . ul . Type2NdefOperationsTest . Config config : org . nfctools . mf . ul . Type2NdefOperationsTest . READ_EMPTY_TEST ) { init ( config ) ; java . util . List < org . nfctools . ndef . Record > ndefMessage = ndefOperations . readNdefMessage ( ) ; "<AssertPlaceHolder>" ; } } readNdefMessage ( ) { assertFormatted ( ) ; if ( ( lastReadRecords ) != null ) { return lastReadRecords ; } else { org . nfctools . mf . tlv . TypeLengthValueReader reader = new org . nfctools . mf . tlv . TypeLengthValueReader ( new org . nfctools . tags . TagInputStream ( memoryLayout , readerWriter ) ) ; convertRecords ( reader ) ; return lastReadRecords ; } }
org . junit . Assert . assertEquals ( 0 , ndefMessage . size ( ) )
testEmptyJson ( ) { io . vertx . core . json . JsonObject configuration = new io . vertx . core . json . JsonObject ( ) ; io . strimzi . operator . cluster . model . AbstractConfiguration config = new io . strimzi . operator . cluster . model . TestConfigurationWithoutDefaults ( configuration ) ; "<AssertPlaceHolder>" ; } getConfiguration ( ) { return configuration ; }
org . junit . Assert . assertTrue ( config . getConfiguration ( ) . isEmpty ( ) )
assertExecuteBatchForMultiplePreparedStatementsFailure ( ) { java . sql . PreparedStatement preparedStatement1 = getPreparedStatement ( ) ; java . sql . PreparedStatement preparedStatement2 = getPreparedStatement ( ) ; java . sql . SQLException exp = new java . sql . SQLException ( ) ; when ( preparedStatement1 . executeBatch ( ) ) . thenThrow ( exp ) ; when ( preparedStatement2 . executeBatch ( ) ) . thenThrow ( exp ) ; setExecuteGroups ( java . util . Arrays . asList ( preparedStatement1 , preparedStatement2 ) ) ; "<AssertPlaceHolder>" ; verify ( preparedStatement1 ) . executeBatch ( ) ; verify ( preparedStatement2 ) . executeBatch ( ) ; } executeBatch ( ) { final boolean isExceptionThrown = org . apache . shardingsphere . core . execute . sql . execute . threadlocal . ExecutorExceptionHandler . isExceptionThrown ( ) ; org . apache . shardingsphere . core . execute . sql . execute . SQLExecuteCallback < int [ ] > callback = new org . apache . shardingsphere . core . execute . sql . execute . SQLExecuteCallback < int [ ] > ( getDatabaseType ( ) , isExceptionThrown ) { @ org . apache . shardingsphere . shardingjdbc . executor . Override protected int [ ] executeSQL ( final org . apache . shardingsphere . core . route . RouteUnit routeUnit , final java . sql . Statement statement , final org . apache . shardingsphere . core . constant . ConnectionMode connectionMode ) throws java . sql . SQLException { return statement . executeBatch ( ) ; } } ; java . util . List < int [ ] > results = executeCallback ( callback ) ; if ( isAccumulate ( ) ) { return accumulate ( results ) ; } else { return results . get ( 0 ) ; } }
org . junit . Assert . assertThat ( actual . executeBatch ( ) , org . hamcrest . CoreMatchers . is ( new int [ ] { 0 , 0 } ) )
testWideChar ( ) { java . io . StringWriter sw = new java . io . StringWriter ( ) ; org . jline . utils . WriterOutputStream wos = new org . jline . utils . WriterOutputStream ( sw , java . nio . charset . StandardCharsets . UTF_8 ) ; byte [ ] bytes = "" . getBytes ( StandardCharsets . UTF_8 ) ; for ( byte b : bytes ) { wos . write ( b ) ; } wos . flush ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( "isSet" + ( optSet ) ) + "\nArg" ) + ( optArg ) ) + "\nargs" ) + ( xargs ) ; }
org . junit . Assert . assertEquals ( "" , sw . toString ( ) )
test41 ( ) { net . pms . formats . SubRip subRip0 = new net . pms . formats . SubRip ( ) ; java . lang . String [ ] str_array1 = subRip0 . getSupportedExtensions ( ) ; "<AssertPlaceHolder>" ; } getSupportedExtensions ( ) { return new java . lang . String [ ] { "ogg" , "ogm" , "ogv" } ; }
org . junit . Assert . assertNotNull ( str_array1 )
queryAlertsCombined ( ) { int numTenants = 1 ; int numTriggers = 5 ; int numAlerts = 5 ; createTestAlerts ( numTenants , numTriggers , numAlerts ) ; java . util . Set < java . lang . String > tenantIds = new java . util . HashSet ( ) ; tenantIds . add ( "tenant0" ) ; org . hawkular . alerts . api . services . AlertsCriteria criteria = new org . hawkular . alerts . api . services . AlertsCriteria ( ) ; criteria . setStatus ( Alert . Status . RESOLVED ) ; criteria . setTriggerId ( "trigger0" ) ; criteria . setStartTime ( 3L ) ; java . util . List < org . hawkular . alerts . api . model . event . Alert > resolvedAlerts = alerts . getAlerts ( tenantIds , criteria , null ) ; "<AssertPlaceHolder>" ; deleteTestAlerts ( numTenants ) ; } size ( ) { return wrapped . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , resolvedAlerts . size ( ) )
parseTrue ( ) { java . lang . Boolean expected = true ; io . robe . convert . xml . parsers . ParseBool parseBool = new io . robe . convert . xml . parsers . ParseBool ( ) ; java . lang . Boolean actual = parseBool . parse ( io . robe . convert . xml . ParserUtil . getParser ( "<bool>true</bool>" ) , null ) ; "<AssertPlaceHolder>" ; } getParser ( java . lang . String ) { com . fasterxml . jackson . dataformat . xml . XmlFactory factory = new com . fasterxml . jackson . dataformat . xml . XmlFactory ( ) ; com . fasterxml . jackson . core . JsonParser parser = factory . createParser ( xml ) ; parser . nextToken ( ) ; parser . nextToken ( ) ; parser . nextToken ( ) ; return parser ; }
org . junit . Assert . assertEquals ( expected , actual )
ExtensionTest ( ) { java . lang . ClassLoader cl = this . getClass ( ) . getClassLoader ( ) ; java . io . File queryFile = new java . io . File ( cl . getResource ( "ParserTest/extension_queries.txt" ) . getFile ( ) ) ; test . robustopt . Map < java . lang . String , test . robustopt . Set < java . lang . String > > schemas = new test . robustopt . HashMap < java . lang . String , test . robustopt . Set < java . lang . String > > ( ) ; test . robustopt . Set < java . lang . String > sailors = new test . robustopt . HashSet < java . lang . String > ( test . robustopt . Arrays . asList ( "sid" , "sname" , "rday" 1 , "age" ) ) ; test . robustopt . Set < java . lang . String > boats = new test . robustopt . HashSet < java . lang . String > ( test . robustopt . Arrays . asList ( "bid" , "rday" 2 , "color" ) ) ; test . robustopt . Set < java . lang . String > reserves = new test . robustopt . HashSet < java . lang . String > ( test . robustopt . Arrays . asList ( "sid" , "bid" , "rday" ) ) ; schemas . put ( "Sailors" , sailors ) ; schemas . put ( "rday" 0 , boats ) ; schemas . put ( "Reserves" , reserves ) ; edu . umich . robustopt . staticanalysis . SQLQueryAnalyzer . setConfig ( new edu . umich . robustopt . staticanalysis . SQLQueryAnalyzer . Configuration ( 'u' ) ) ; edu . umich . robustopt . staticanalysis . SQLQueryAnalyzer analyzer = new edu . umich . robustopt . staticanalysis . SQLQueryAnalyzer ( ) ; analyzer . setVerbose ( true ) ; analyzer . analyzeFile ( queryFile , schemas ) ; "<AssertPlaceHolder>" ; System . out . println ( "extension<sp>stats<sp>test<sp>passed." ) ; } hasUnresolvedColumn ( ) { return analyzer . hasUnresolvedSymbol ( ) ; }
org . junit . Assert . assertFalse ( analyzer . hasUnresolvedColumn ( ) )
listAllAccessLevelsForwardsCalls ( ) { final java . util . List < org . phenotips . data . permissions . AccessLevel > levels = java . util . Arrays . asList ( org . phenotips . data . permissions . internal . DefaultEntityPermissionsManagerTest . VIEW_ACCESS , org . phenotips . data . permissions . internal . DefaultEntityPermissionsManagerTest . EDIT_ACCESS ) ; when ( this . accessManager . listAllAccessLevels ( ) ) . thenReturn ( levels ) ; "<AssertPlaceHolder>" ; verify ( this . accessManager , times ( 1 ) ) . listAllAccessLevels ( ) ; } listAllAccessLevels ( ) { return this . internalService . listAllAccessLevels ( ) ; }
org . junit . Assert . assertSame ( levels , this . component . listAllAccessLevels ( ) )
testGetQueryGTEFilters ( ) { filters . add ( org . nohope . cassandra . mapservice . cfilter . CFilters . gte ( org . nohope . cassandra . mapservice . Value . bound ( org . nohope . cassandra . mapservice . CFilterBuilderTest . INT_COL , 4 ) ) ) ; final java . util . List < org . nohope . cassandra . mapservice . cfilter . CFilter < ? > > builderFilters = org . nohope . cassandra . mapservice . CFilterBuilder . getQueryFilters ( ) . gte ( org . nohope . cassandra . mapservice . CFilterBuilderTest . INT_COL , 4 ) . getFilters ( ) ; "<AssertPlaceHolder>" ; } getFilters ( ) { return java . util . Collections . unmodifiableCollection ( filters ) ; }
org . junit . Assert . assertEquals ( filters , builderFilters )
testIsValidWhenDataTypeNameIsBlank ( ) { final org . kie . workbench . common . dmn . client . editors . types . common . DataType dataType = mock ( org . kie . workbench . common . dmn . client . editors . types . common . DataType . class ) ; final org . kie . workbench . common . dmn . client . editors . common . messages . FlashMessage blankMessage = mock ( org . kie . workbench . common . dmn . client . editors . common . messages . FlashMessage . class ) ; doReturn ( true ) . when ( validator ) . isBlank ( dataType ) ; doReturn ( true ) . when ( validator ) . isNotUnique ( dataType ) ; when ( blankErrorMessage . getFlashMessage ( dataType ) ) . thenReturn ( blankMessage ) ; final boolean isValid = validator . isValid ( dataType ) ; verify ( flashMessageEvent ) . fire ( blankMessage ) ; "<AssertPlaceHolder>" ; } fire ( org . uberfire . workbench . events . NotificationEvent ) { }
org . junit . Assert . assertFalse ( isValid )
testSecondSerialize ( ) { final java . util . Map < java . lang . String , java . lang . Object > m = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; m . put ( "three" , 3 ) ; m . put ( "hello" , "world" ) ; m . put ( "empty" , new java . util . HashMap < java . lang . Object , java . lang . Object > ( ) ) ; final com . lightboxtechnologies . spectrum . JsonWritable w = new com . lightboxtechnologies . spectrum . JsonWritable ( ) ; w . set ( m ) ; "<AssertPlaceHolder>" ; } toString ( ) { final org . apache . commons . codec . binary . Hex hex = new org . apache . commons . codec . binary . Hex ( ) ; final byte [ ] bytes = new byte [ getLength ( ) ] ; java . lang . System . arraycopy ( getBytes ( ) , 0 , bytes , 0 , bytes . length ) ; return new java . lang . String ( hex . encode ( bytes ) ) ; }
org . junit . Assert . assertEquals ( w . toString ( ) , w . toString ( ) )
lazylift_adds_chained_maps_support ( ) { java . util . Collection < java . net . URI > result = uris . stream ( ) . map ( lazy ( URI :: create ) ) . map ( lazylift ( ( uri ) -> { if ( uri . toString ( ) . startsWith ( "ftp" ) ) throw new java . lang . UnsupportedOperationException ( ) ; else return uri ; } ) ) . collect ( discarding ( me . hgwood . bulky . IllegalArgumentException . class , me . hgwood . bulky . UnsupportedOperationException . class ) ) . collect ( toList ( ) ) ; "<AssertPlaceHolder>" ; } lazylift ( java . util . function . Function ) { return me . hgwood . bulky . Bulky . liftToSuppliers ( f ) ; }
org . junit . Assert . assertThat ( result , org . hamcrest . Matchers . contains ( java . net . URI . create ( uris . get ( 0 ) ) , java . net . URI . create ( uris . get ( 3 ) ) ) )
shouldRaiseNoConstraintViolationOnCalculateVAT ( ) { org . agoncal . book . javaee7 . chapter03 . ex10 . Item10 item = new org . agoncal . book . javaee7 . chapter03 . ex10 . Item10 ( "H2G2" , 12.5F , "Best<sp>IT<sp>Scifi<sp>Book" ) ; javax . validation . executable . ExecutableValidator methodValidator = org . agoncal . book . javaee7 . chapter03 . ex10 . Item10Test . validator . forExecutables ( ) ; java . lang . reflect . Method method = org . agoncal . book . javaee7 . chapter03 . ex10 . Item10 . class . getMethod ( "calculateVAT" ) ; java . util . Set < javax . validation . ConstraintViolation < org . agoncal . book . javaee7 . chapter03 . ex10 . Item10 > > violations = methodValidator . validateReturnValue ( item , method , new java . lang . Float ( 10.0 ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( 0 , violations . size ( ) )
propertiesLeadingSpaceAfterNewLine ( ) { java . util . Map < java . lang . String , java . lang . String > actual = io . strimzi . operator . cluster . model . OrderedPropertiesTest . propertiesCompatibility ( "key<sp>:<sp>multi\\\n<sp>\\\r\n<sp>\t<sp>line" ) ; io . strimzi . operator . cluster . model . OrderedProperties expected = new io . strimzi . operator . cluster . model . OrderedProperties ( ) . addPair ( "key" , "multiline" ) ; "<AssertPlaceHolder>" ; } asMap ( ) { return pairs ; }
org . junit . Assert . assertEquals ( expected . asMap ( ) , actual )
appendJsonAttribute_good_case_value_null ( ) { java . lang . StringBuilder stringBuilder = new java . lang . StringBuilder ( ) ; stringBuilder . append ( "prefix_" ) ; java . lang . String expResult = "prefix_\"keyname\":null," ; com . microsoft . azure . sdk . iot . provisioning . service . Tools . appendJsonAttribute ( stringBuilder , "keyname" , null , false , false ) ; "<AssertPlaceHolder>" ; } toString ( ) { com . google . gson . Gson gson = new com . google . gson . GsonBuilder ( ) . disableHtmlEscaping ( ) . setPrettyPrinting ( ) . create ( ) ; return gson . toJson ( this ) ; }
org . junit . Assert . assertEquals ( expResult , stringBuilder . toString ( ) )
testParameter ( ) { final net . violet . platform . api . actions . Action theAction = new net . violet . platform . api . actions . stores . Delete ( ) ; final net . violet . platform . api . callers . APICaller caller = getPublicApplicationAPICaller ( ) ; final java . util . Map < java . lang . String , java . lang . Object > theParams = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; theParams . put ( "hardware" , Hardware . HARDWARE . DALDAL ) ; final net . violet . platform . api . actions . ActionParam theActionParam = new net . violet . platform . api . actions . ActionParam ( caller , theParams ) ; final java . lang . Object theResult = theAction . processRequest ( theActionParam ) ; "<AssertPlaceHolder>" ; } put ( K , T ) { this . mMap . put ( theRef , new net . violet . db . cache . CacheReference < K , T > ( theRef , theRecord , this . mReferenceQueue ) ) ; this . mLinkedMap . put ( theRef , theRecord ) ; }
org . junit . Assert . assertNotNull ( theResult )
distinct_Request ( ) { java . lang . String [ ] expected = new java . lang . String [ ] { "v1" , "v3" , "v4" } ; java . util . List < java . lang . String > actual = com . careerbuilder . search . relevancy . utility . ResponseUtility . distinct ( requestValuesDups ) ; "<AssertPlaceHolder>" ; } distinct ( java . lang . String [ ] ) { return new com . careerbuilder . search . relevancy . utility . LinkedList ( com . careerbuilder . search . relevancy . utility . Arrays . stream ( ( requestValues == null ? new java . lang . String [ 0 ] : requestValues ) ) . collect ( java . util . stream . Collectors . toMap ( String :: toLowerCase , ( s ) -> s , ( first , second ) -> first ) ) . values ( ) ) ; }
org . junit . Assert . assertArrayEquals ( expected , actual . toArray ( new java . lang . String [ 0 ] ) )
testQ1Y1900Previous ( ) { org . jfree . data . time . Quarter previous = ( ( org . jfree . data . time . Quarter ) ( this . q1Y1900 . previous ( ) ) ) ; "<AssertPlaceHolder>" ; } previous ( ) { org . jfree . data . time . Month result ; if ( ( this . month ) != ( org . jfree . date . MonthConstants . JANUARY ) ) { result = new org . jfree . data . time . Month ( ( ( this . month ) - 1 ) , this . year ) ; } else { if ( ( this . year ) > 1900 ) { result = new org . jfree . data . time . Month ( org . jfree . date . MonthConstants . DECEMBER , ( ( this . year ) - 1 ) ) ; } else { result = null ; } } return result ; }
org . junit . Assert . assertNull ( previous )
shouldUpdateHttpProxy ( ) { verifyProxy ( Configuration . INSTANCE . getHttpProxySettings ( ) , false , "host-http" , 9000 , "user-http" , "password-http" ) ; byte [ ] initFileContents = org . apache . commons . io . FileUtils . readFileToByteArray ( configSettingFile ) ; boolean activated = true ; java . lang . String host = "host.of.the.new.proxy" ; java . lang . String password = "the*new_password" ; java . lang . String username = "the<sp>new<sp>user<sp>name" ; int port = 8567 ; io . uploader . drive . config . proxy . Proxy newProxy = new io . uploader . drive . config . proxy . Proxy . Builder ( "http" ) . setActivated ( activated ) . setHost ( host ) . setPassword ( password ) . setUsername ( username ) . setPort ( port ) . build ( ) ; Configuration . INSTANCE . updateProxy ( newProxy ) ; verifyProxy ( Configuration . INSTANCE . getHttpProxySettings ( ) , activated , host , port , username , password ) ; byte [ ] updatedFileContents = org . apache . commons . io . FileUtils . readFileToByteArray ( configSettingFile ) ; "<AssertPlaceHolder>" ; } getHttpProxySettings ( ) { return httpProxySettings ; }
org . junit . Assert . assertFalse ( java . util . Arrays . equals ( updatedFileContents , initFileContents ) )
testCfsettingNoAttributes ( ) { final java . lang . String cfcSrc = "<cfset<sp>var<sp>a<sp>=<sp>1>" ; com . cflint . api . CFLintResult lintresult = cfBugs . scan ( cfcSrc , "test" ) ; "<AssertPlaceHolder>" ; } getIssues ( ) { return cflint . getBugs ( ) . getBugList ( ) ; }
org . junit . Assert . assertEquals ( 0 , lintresult . getIssues ( ) . size ( ) )
testFetchByPrimaryKeyExisting ( ) { com . liferay . site . navigation . model . SiteNavigationMenuItem newSiteNavigationMenuItem = addSiteNavigationMenuItem ( ) ; com . liferay . site . navigation . model . SiteNavigationMenuItem existingSiteNavigationMenuItem = _persistence . fetchByPrimaryKey ( newSiteNavigationMenuItem . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
org . junit . Assert . assertEquals ( existingSiteNavigationMenuItem , newSiteNavigationMenuItem )
sendMessage_withJMSResources ( ) { org . mockito . MockitoAnnotations . initMocks ( this ) ; doReturn ( Boolean . TRUE ) . when ( sender ) . checkJMSResources ( ) ; javax . jms . Connection conn = mock ( javax . jms . Connection . class ) ; doReturn ( conn ) . when ( factory ) . createConnection ( ) ; javax . jms . Session session = mock ( javax . jms . Session . class ) ; doReturn ( session ) . when ( conn ) . createSession ( eq ( false ) , eq ( Session . AUTO_ACKNOWLEDGE ) ) ; javax . jms . MessageProducer producer = mock ( javax . jms . MessageProducer . class ) ; doReturn ( producer ) . when ( session ) . createProducer ( eq ( queue ) ) ; javax . jms . ObjectMessage om = mock ( javax . jms . ObjectMessage . class ) ; doReturn ( om ) . when ( session ) . createObjectMessage ( ) ; doNothing ( ) . when ( om ) . setObject ( caughtMessage . capture ( ) ) ; sender . sendMessage ( objectMessage ) ; verify ( factory , times ( 1 ) ) . createConnection ( ) ; verify ( session , times ( 1 ) ) . close ( ) ; verify ( conn , times ( 1 ) ) . close ( ) ; verify ( producer , times ( 1 ) ) . send ( eq ( om ) ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return value ; }
org . junit . Assert . assertEquals ( objectMessage , caughtMessage . getValue ( ) )
testSerialization ( ) { org . jfree . chart . renderer . xy . StandardXYBarPainter p1 = new org . jfree . chart . renderer . xy . StandardXYBarPainter ( ) ; org . jfree . chart . renderer . xy . StandardXYBarPainter p2 = ( ( org . jfree . chart . renderer . xy . StandardXYBarPainter ) ( org . jfree . chart . TestUtils . serialised ( p1 ) ) ) ; "<AssertPlaceHolder>" ; } serialised ( java . lang . Object ) { java . lang . Object result = null ; java . io . ByteArrayOutputStream buffer = new java . io . ByteArrayOutputStream ( ) ; java . io . ObjectOutput out ; try { out = new java . io . ObjectOutputStream ( buffer ) ; out . writeObject ( original ) ; out . close ( ) ; java . io . ObjectInput in = new java . io . ObjectInputStream ( new java . io . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; result = in . readObject ( ) ; in . close ( ) ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } catch ( java . lang . ClassNotFoundException e ) { throw new java . lang . RuntimeException ( e ) ; } return result ; }
org . junit . Assert . assertEquals ( p1 , p2 )
testMiniCluster ( ) { org . apache . hadoop . hbase . HBaseTestingUtility hbt = new org . apache . hadoop . hbase . HBaseTestingUtility ( ) ; org . apache . hadoop . hbase . MiniHBaseCluster cluster = hbt . startMiniCluster ( ) ; try { "<AssertPlaceHolder>" ; } finally { hbt . shutdownMiniCluster ( ) ; } } getLiveRegionServerThreads ( ) { return this . hbaseCluster . getLiveRegionServers ( ) ; }
org . junit . Assert . assertEquals ( 1 , cluster . getLiveRegionServerThreads ( ) . size ( ) )
bootstrap ( ) { com . buschmais . xo . api . XOManager xoManager = getXOManager ( ) ; xoManager . currentTransaction ( ) . begin ( ) ; com . puresoltechnologies . ductiledb . xo . test . bootstrap . TestEntity a = xoManager . create ( com . puresoltechnologies . ductiledb . xo . test . bootstrap . TestEntity . class ) ; a . setName ( "Test" ) ; xoManager . currentTransaction ( ) . commit ( ) ; xoManager . currentTransaction ( ) . begin ( ) ; com . buschmais . xo . api . Query < com . puresoltechnologies . ductiledb . xo . test . bootstrap . TestEntity > query = xoManager . createQuery ( "g.V().has('name','Test')" , com . puresoltechnologies . ductiledb . xo . test . bootstrap . TestEntity . class ) ; com . puresoltechnologies . ductiledb . xo . test . bootstrap . TestEntity readA = query . execute ( ) . getSingleResult ( ) ; "<AssertPlaceHolder>" ; xoManager . currentTransaction ( ) . commit ( ) ; } getName ( ) { return name ; }
org . junit . Assert . assertEquals ( a . getName ( ) , readA . getName ( ) )
getCanonicalURI_invalid_no_host2 ( ) { java . net . URI result = org . kaazing . gateway . transport . http . HttpUtils . getCanonicalURI ( "ws://:8026" , true ) ; "<AssertPlaceHolder>" ; } getHost ( ) { return host ; }
org . junit . Assert . assertNull ( result . getHost ( ) )
testValidIdWithDashes ( ) { java . lang . String id = "634e8505-bd4b-436e-97e8-2045d1b0d265" ; java . lang . String qualifier = "zoom-and-enhanced-overview" ; ddf . catalog . content . data . ContentItem item = new ddf . catalog . content . data . impl . ContentItemImpl ( id , qualifier , null , "" , null ) ; "<AssertPlaceHolder>" ; } validate ( ddf . catalog . content . data . ContentItem ) { if ( ( item == null ) || ( org . apache . commons . lang . StringUtils . isBlank ( item . getUri ( ) ) ) ) { return false ; } if ( org . apache . commons . lang . StringUtils . isNotBlank ( item . getQualifier ( ) ) ) { boolean qualifierValid = ddf . catalog . content . data . impl . ContentItemValidator . validateInput ( item . getQualifier ( ) , ddf . catalog . content . data . impl . ContentItemValidator . QUALIFIER_PATTERN ) ; if ( ! qualifierValid ) { return false ; } } if ( ddf . catalog . content . data . impl . ContentItemValidator . CONTENT_PATTERN . matcher ( item . getUri ( ) ) . matches ( ) ) { return true ; } return false ; }
org . junit . Assert . assertThat ( ddf . catalog . content . data . impl . ContentItemValidator . validate ( item ) , org . hamcrest . Matchers . is ( true ) )
testGetBuilderIdempotent ( ) { setUp ( ) ; org . apache . olingo . odata2 . jpa . processor . api . access . JPAEdmBuilder builder1 = objJPAEdmProperty . getBuilder ( ) ; org . apache . olingo . odata2 . jpa . processor . api . access . JPAEdmBuilder builder2 = objJPAEdmProperty . getBuilder ( ) ; "<AssertPlaceHolder>" ; } hashCode ( ) { final int prime = 31 ; int result = 1 ; result = ( prime * result ) + ( ( key ) == null ? 0 : key . hashCode ( ) ) ; return result ; }
org . junit . Assert . assertEquals ( builder1 . hashCode ( ) , builder2 . hashCode ( ) )
testMergeNoFF ( ) { setMerge ( FastForwardMode . NO_FF ) ; org . eclipse . egit . core . op . MergeOperation operation = new org . eclipse . egit . core . op . MergeOperation ( testRepository . getRepository ( ) , org . eclipse . egit . core . test . op . MergeOperationTest . MASTER ) ; operation . execute ( new org . eclipse . core . runtime . NullProgressMonitor ( ) ) ; "<AssertPlaceHolder>" ; } countCommitsInHead ( ) { org . eclipse . jgit . api . LogCommand log ; try ( org . eclipse . jgit . api . Git git = new org . eclipse . jgit . api . Git ( testRepository . getRepository ( ) ) ) { log = git . log ( ) ; } java . lang . Iterable < org . eclipse . jgit . revwalk . RevCommit > commits = log . call ( ) ; int result = 0 ; for ( java . util . Iterator i = commits . iterator ( ) ; i . hasNext ( ) ; ) { i . next ( ) ; result ++ ; } return result ; }
org . junit . Assert . assertEquals ( 3 , countCommitsInHead ( ) )
shouldUninstallNotTemporaryPluginFile ( ) { java . io . File pluginFile = new java . io . File ( destination , "uninstallpluginname.jar" ) ; org . apache . commons . io . FileUtils . touch ( pluginFile ) ; defaultPluginFileManager . uninstallPlugin ( "uninstallpluginname.jar" ) ; "<AssertPlaceHolder>" ; } uninstallPlugin ( java . lang . String [ ] ) { for ( java . lang . String filename : filenames ) { java . io . File file = new java . io . File ( ( ( ( pluginsTmpPath ) + ( java . lang . System . getProperty ( com . qcadoo . plugin . internal . filemanager . DefaultPluginFileManager . L_FILE_SEPARATOR ) ) ) + filename ) ) ; if ( ! ( file . exists ( ) ) ) { file = new java . io . File ( ( ( ( pluginsPath ) + ( java . lang . System . getProperty ( com . qcadoo . plugin . internal . filemanager . DefaultPluginFileManager . L_FILE_SEPARATOR ) ) ) + filename ) ) ; } try { org . apache . commons . io . FileUtils . forceDelete ( file ) ; } catch ( java . io . IOException e ) { com . qcadoo . plugin . internal . filemanager . DefaultPluginFileManager . LOG . error ( ( "Problem<sp>with<sp>removing<sp>plugin<sp>file<sp>-<sp>" + ( e . getMessage ( ) ) ) ) ; if ( file . exists ( ) ) { com . qcadoo . plugin . internal . filemanager . DefaultPluginFileManager . LOG . info ( "Trying<sp>delete<sp>file<sp>after<sp>JVM<sp>stop" ) ; file . deleteOnExit ( ) ; } } } }
org . junit . Assert . assertFalse ( pluginFile . exists ( ) )
testPropertyIsLikeMultipleTermsWithWildcard ( ) { java . lang . String searchPhrase = "abc<sp>123*" ; java . lang . String expectedQuery = ( SolrFilterDelegate . WHITESPACE_TOKENIZED_METADATA_FIELD ) + ":(abc<sp>123*)" ; org . apache . solr . client . solrj . SolrQuery isLikeQuery = toTest . propertyIsLike ( Metacard . ANY_TEXT , searchPhrase , false ) ; "<AssertPlaceHolder>" ; } getQuery ( ) { return query ; }
org . junit . Assert . assertThat ( isLikeQuery . getQuery ( ) , org . hamcrest . CoreMatchers . is ( expectedQuery ) )
testDoNothingOneElementConfig ( ) { org . dcache . gplazma . strategies . MappingStrategy strategy = strategyFactory . newMappingStrategy ( ) ; "<AssertPlaceHolder>" ; strategy . setPlugins ( oneDoNothingPlugins ) ; java . util . Set < java . security . Principal > principals = com . google . common . collect . Sets . newHashSet ( ) ; strategy . map ( org . dcache . gplazma . strategies . MappingStrategyMapTests . IGNORING_LOGIN_MONITOR , principals ) ; } newMappingStrategy ( ) { return new org . dcache . gplazma . strategies . DefaultMappingStrategy ( ) ; }
org . junit . Assert . assertNotNull ( strategy )
testIntParseFailure ( ) { java . lang . Integer num = com . rometools . rome . io . impl . NumberParser . parseInt ( "Non<sp>Int" ) ; "<AssertPlaceHolder>" ; } parseInt ( java . lang . String ) { if ( null != str ) { try { return new java . lang . Integer ( java . lang . Integer . parseInt ( str . trim ( ) ) ) ; } catch ( final java . lang . Exception e ) { } } return null ; }
org . junit . Assert . assertNull ( num )
setLoggedOffShouldHandleMissingController ( ) { "<AssertPlaceHolder>" ; chatWindow . setLoggedOff ( ) ; verifyZeroInteractions ( controller ) ; } getControllerFromChatWindow ( ) { return net . usikkert . kouchat . util . TestUtils . getFieldValue ( chatWindow , net . usikkert . kouchat . android . controller . PrivateChatController . class , "privateChatController" ) ; }
org . junit . Assert . assertNull ( getControllerFromChatWindow ( ) )
testBuildTableNameNonRootServicePathDataModelByEntityEncoding ( ) { System . out . println ( ( ( ( ( getTestTraceHead ( "someId=someType" 4 ) ) + "--------<sp>When<sp>encoding<sp>and<sp>when<sp>a<sp>non<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>and<sp>data_model<sp>is<sp>" ) + "'dm-by-service-path'<sp>the<sp>MySQL<sp>table<sp>name<sp>is<sp>the<sp>encoding<sp>of<sp>the<sp>concatenation<sp>of<sp><service-path>,<sp>" ) + "someId=someType" 1 ) ) ; java . lang . String attrPersistence = null ; java . lang . String batchSize = null ; java . lang . String batchTime = null ; java . lang . String batchTTL = null ; java . lang . String dataModel = "dm-by-entity" ; java . lang . String enableEncoding = "true" ; java . lang . String enableGrouping = null ; java . lang . String enableLowercase = null ; java . lang . String host = null ; java . lang . String password = null ; java . lang . String port = null ; java . lang . String username = null ; com . telefonica . iot . cygnus . sinks . NGSIMySQLSink sink = new com . telefonica . iot . cygnus . sinks . NGSIMySQLSink ( ) ; sink . configure ( createContext ( attrPersistence , batchSize , batchTime , batchTTL , dataModel , enableEncoding , enableGrouping , enableLowercase , host , password , port , username ) ) ; java . lang . String servicePath = "/somePath" ; java . lang . String entity = "someId=someType" ; java . lang . String attribute = null ; try { java . lang . String builtTableName = sink . buildTableName ( servicePath , entity , attribute ) ; java . lang . String expecetedTableName = "x002fsomePathxffffsomeIdxffffsomeType" ; try { "<AssertPlaceHolder>" ; System . out . println ( ( ( ( ( ( getTestTraceHead ( "someId=someType" 4 ) ) + "-<sp>OK<sp>-<sp>'" ) + builtTableName ) + "'<sp>is<sp>equals<sp>to<sp>the<sp>encoding<sp>of<sp><service-path>,<sp><entityId><sp>" ) + "someId=someType" 2 ) ) ; } catch ( java . lang . AssertionError e ) { System . out . println ( ( ( ( ( ( getTestTraceHead ( "someId=someType" 4 ) ) + "-<sp>FAIL<sp>-<sp>'" ) + builtTableName ) + "someId=someType" 0 ) + "someId=someType" 1 ) ) ; throw e ; } } catch ( java . lang . Exception e ) { System . out . println ( ( ( getTestTraceHead ( "someId=someType" 4 ) ) + "someId=someType" 3 ) ) ; throw e ; } buildTableName ( java . lang . String , java . lang . String , java . lang . String ) { java . lang . String name ; switch ( dataModel ) { case DMBYSERVICEPATH : name = com . telefonica . iot . cygnus . utils . NGSICharsets . encodePostgreSQL ( servicePath ) ; break ; case DMBYENTITY : java . lang . String truncatedServicePath = com . telefonica . iot . cygnus . utils . NGSICharsets . encodePostgreSQL ( servicePath ) ; name = ( truncatedServicePath . isEmpty ( ) ? "" : truncatedServicePath + ( com . telefonica . iot . cygnus . utils . CommonConstants . CONCATENATOR ) ) + ( com . telefonica . iot . cygnus . utils . NGSICharsets . encodePostgreSQL ( entity ) ) ; break ; case DMBYATTRIBUTE : truncatedServicePath = com . telefonica . iot . cygnus . utils . NGSICharsets . encodePostgreSQL ( servicePath ) ; name = ( ( ( truncatedServicePath . isEmpty ( ) ? "" : truncatedServicePath + ( com . telefonica . iot . cygnus . utils . CommonConstants . CONCATENATOR ) ) + ( com . telefonica . iot . cygnus . utils . NGSICharsets . encodePostgreSQL ( entity ) ) ) + ( com . telefonica . iot . cygnus . utils . CommonConstants . CONCATENATOR ) ) + ( com . telefonica . iot . cygnus . utils . NGSICharsets . encodePostgreSQL ( attribute ) ) ; break ; default : throw new com . telefonica . iot . cygnus . errors . CygnusBadConfiguration ( ( ( "Unknown<sp>data<sp>model<sp>'" + ( dataModel . toString ( ) ) ) + "'.<sp>Please,<sp>use<sp>dm-by-service-path,<sp>dm-by-entity<sp>or<sp>dm-by-attribute" ) ) ; } if ( ( name . length ( ) ) > ( com . telefonica . iot . cygnus . utils . NGSIConstants . POSTGRESQL_MAX_NAME_LEN ) ) { throw new com . telefonica . iot . cygnus . errors . CygnusBadConfiguration ( ( ( ( "Building<sp>table<sp>name<sp>'" + name ) + "'<sp>and<sp>its<sp>length<sp>is<sp>greater<sp>than<sp>" ) + ( com . telefonica . iot . cygnus . utils . NGSIConstants . POSTGRESQL_MAX_NAME_LEN ) ) ) ; } return name ; }
org . junit . Assert . assertEquals ( expecetedTableName , builtTableName )
testCustomLDPath2 ( ) { java . util . List < com . github . anno4j . model . Annotation > list = queryService . addCriteria ( "fn:getSelector(.)[is-a<sp>oa:FragmentSelector]" ) . execute ( ) ; "<AssertPlaceHolder>" ; } size ( ) { try { if ( ( _size ) < 0 ) { synchronized ( this ) { if ( ( _size ) < 0 ) { int index = findSize ( ) ; _size = index ; } } } return _size ; } catch ( org . openrdf . repository . RepositoryException e ) { throw new org . openrdf . repository . object . exceptions . ObjectStoreException ( e ) ; } }
org . junit . Assert . assertEquals ( 1 , list . size ( ) )
noSuchUser ( ) { fetchOrCreateUser ( ) ; org . apache . shiro . authc . UsernamePasswordToken upToken = new org . apache . shiro . authc . UsernamePasswordToken ( "NoSuchUser" , getPlainTextPassword ( ) ) ; org . apache . shiro . authc . AuthenticationInfo authNInfo = io . ifar . security . realm . TestJdbiShiroRealm . realm . doGetAuthenticationInfo ( upToken ) ; "<AssertPlaceHolder>" ; } doGetAuthenticationInfo ( io . ifar . security . realm . AuthenticationToken ) { checkArgument ( io . ifar . security . realm . UsernamePasswordToken . class . isAssignableFrom ( authNToken . getClass ( ) ) , "doGetAuthenticationInfo():<sp>AuthenticationToken<sp>argument<sp>needs<sp>to<sp>be<sp>an<sp>instance<sp>of<sp>UsernamePasswordToken.<sp>Was<sp>an<sp>instance<sp>of<sp>'%s'<sp>instead." , authNToken . getClass ( ) . getName ( ) ) ; io . ifar . security . realm . UsernamePasswordToken upToken = ( ( io . ifar . security . realm . UsernamePasswordToken ) ( authNToken ) ) ; java . lang . String username = upToken . getUsername ( ) ; if ( com . google . common . base . Strings . isNullOrEmpty ( username ) ) { io . ifar . security . realm . JdbiShiroRealm . LOG . error ( "doGetAuthenticationInfo()<sp>requires<sp>a<sp>non-null,<sp>non-empty<sp>username" ) ; throw new io . ifar . security . realm . AccountException ( "username<sp>is<sp>required<sp>by<sp>this<sp>realm." ) ; } io . ifar . security . realm . model . ISecurityUser user ; try { user = getUserSecurityDAO ( ) . findUserWithoutRoles ( username ) ; } catch ( java . lang . RuntimeException ex ) { io . ifar . security . realm . JdbiShiroRealm . LOG . error ( "Error<sp>retrieving<sp>user<sp>'{}'<sp>from<sp>database.<sp>{}" , username , ex . getMessage ( ) ) ; if ( ex instanceof io . ifar . security . realm . AuthenticationException ) { throw ex ; } else { throw new io . ifar . security . realm . AuthenticationException ( ( ( "Database<sp>is<sp>inconsistent.<sp>Queried<sp>for<sp>user<sp>with<sp>username<sp>of<sp>'{}',<sp>retrieved<sp>username<sp>of<sp>'{}'." 0 + username ) + "'." ) , ex ) ; } } if ( user != null ) { if ( ! ( username . equals ( user . getUsername ( ) ) ) ) { io . ifar . security . realm . JdbiShiroRealm . LOG . error ( "Database<sp>is<sp>inconsistent.<sp>Queried<sp>for<sp>user<sp>with<sp>username<sp>of<sp>'{}',<sp>retrieved<sp>username<sp>of<sp>'{}'." , username , user . getUsername ( ) ) ; throw new io . ifar . security . realm . AccountException ( "database<sp>error:<sp>username<sp>mis-match" ) ; } java . lang . String password = user . getPassword ( ) ; if ( ( isPasswordRequired ( ) ) && ( password == null ) ) { io . ifar . security . realm . JdbiShiroRealm . LOG . warn ( "Password<sp>is<sp>required<sp>and<sp>username<sp>'{}'<sp>has<sp>a<sp>null<sp>password.<sp>Treating<sp>account<sp>as<sp>disabled." , username ) ; throw new io . ifar . security . realm . DisabledAccountException ( ( ( "No<sp>valid<sp>account<sp>found<sp>for<sp>user<sp>'" + username ) + "'." ) ) ; } didAuthentication = true ; java . util . List < io . ifar . security . realm . JdbiShiroRealm . PrincipalValueField > pvs = getPrincipalValueFields ( ) ; java . util . Set < java . lang . Object > principalVals = new java . util . LinkedHashSet ( pvs . size ( ) ) ; for ( io . ifar . security . realm . JdbiShiroRealm . PrincipalValueField pv : pvs ) { if ( io . ifar . security . realm . JdbiShiroRealm . PrincipalValueField . USER_ID . equals ( pv ) ) { principalVals . add ( user . getId ( ) ) ; } else if ( io . ifar . security . realm . JdbiShiroRealm . PrincipalValueField . USERNAME . equals ( pv ) ) { principalVals . add ( user . getUsername ( ) ) ; } } org . apache . shiro . subject . SimplePrincipalCollection spc = new org . apache . shiro . subject . SimplePrincipalCollection ( principalVals , getName ( ) ) ; io . ifar . security . realm . JdbiShiroRealm . LOG . debug ( "Found<sp>user<sp>record.<sp>Returning<sp>authentication<sp>info<sp>with<sp>principal<sp>collection<sp>of:<sp>{}" , spc ) ; return new io . ifar . security . realm . SimpleAuthenticationInfo ( spc , password ) ; } else { return null ; } }
org . junit . Assert . assertNull ( authNInfo )
testMatch ( ) { final io . norberg . rut . Route r = io . norberg . rut . Route . of ( "GET" , "foo" ) ; "<AssertPlaceHolder>" ; } captureNames ( ) { return path . captureNames ( ) ; }
org . junit . Assert . assertThat ( r . captureNames ( ) , org . hamcrest . Matchers . is ( org . hamcrest . Matchers . empty ( ) ) )
testWithoutDelete ( ) { final org . apache . oozie . fluentjob . api . action . ActionAttributesBuilder builder = getBuilderInstance ( ) ; for ( final org . apache . oozie . fluentjob . api . action . Delete delete : org . apache . oozie . fluentjob . api . action . TestActionAttributesBuilder . DELETES ) { builder . withDelete ( delete ) ; } builder . withoutDelete ( org . apache . oozie . fluentjob . api . action . TestActionAttributesBuilder . DELETES [ 0 ] ) ; final org . apache . oozie . fluentjob . api . action . ActionAttributes attributes = builder . build ( ) ; final java . util . List < org . apache . oozie . fluentjob . api . action . Delete > expectedDeletes = java . util . Arrays . asList ( org . apache . oozie . fluentjob . api . action . TestActionAttributesBuilder . DELETES ) . subList ( 1 , org . apache . oozie . fluentjob . api . action . TestActionAttributesBuilder . DELETES . length ) ; "<AssertPlaceHolder>" ; } getDeletes ( ) { return attributes . getDeletes ( ) ; }
org . junit . Assert . assertEquals ( expectedDeletes , attributes . getDeletes ( ) )
sortByClassNameTest ( ) { sortElementList . add ( uncertain5 ) ; expectedElementList . add ( app1 ) ; expectedElementList . add ( app2 ) ; expectedElementList . add ( uncertain5 ) ; expectedElementList . add ( app3 ) ; java . util . List < com . dianping . kernel . plugin . SortTool . SortElement > sel = new com . dianping . kernel . plugin . SortTool ( ) . sort ( sortElementList ) ; "<AssertPlaceHolder>" ; } sort ( java . util . List ) { java . util . List < com . dianping . kernel . plugin . SortTool . SortElementWrapper > allWrapperList = new java . util . ArrayList < com . dianping . kernel . plugin . SortTool . SortElementWrapper > ( ) ; java . util . List < com . dianping . kernel . plugin . SortTool . SortElementWrapper > beforeWrapperList = new java . util . ArrayList < com . dianping . kernel . plugin . SortTool . SortElementWrapper > ( ) ; java . util . List < com . dianping . kernel . plugin . SortTool . SortElementWrapper > duringWrapperList = new java . util . ArrayList < com . dianping . kernel . plugin . SortTool . SortElementWrapper > ( ) ; java . util . List < com . dianping . kernel . plugin . SortTool . SortElementWrapper > afterWrapperList = new java . util . ArrayList < com . dianping . kernel . plugin . SortTool . SortElementWrapper > ( ) ; java . util . List < com . dianping . kernel . plugin . SortTool . SortElementWrapper > uncertainWrapperList = new java . util . ArrayList < com . dianping . kernel . plugin . SortTool . SortElementWrapper > ( ) ; java . util . Set < com . dianping . kernel . plugin . SortTool . SortElementWrapper > notMatchWrapperSet = new java . util . HashSet < com . dianping . kernel . plugin . SortTool . SortElementWrapper > ( ) ; if ( ( elementList != null ) && ( ( elementList . size ( ) ) > 0 ) ) { for ( int i = 0 ; i < ( elementList . size ( ) ) ; i ++ ) { com . dianping . kernel . plugin . SortTool . SortElement element = elementList . get ( i ) ; com . dianping . kernel . plugin . SortTool . SortElementWrapper wrapper = new com . dianping . kernel . plugin . SortTool . SortElementWrapper ( i , element ) ; allWrapperList . add ( wrapper ) ; } for ( com . dianping . kernel . plugin . SortTool . SortElementWrapper wrapper : allWrapperList ) { segment ( wrapper . getElement ( ) . getRule ( ) , wrapper , allWrapperList , beforeWrapperList , duringWrapperList , afterWrapperList , uncertainWrapperList , notMatchWrapperSet ) ; } for ( com . dianping . kernel . plugin . SortTool . SortElementWrapper uncertainWrapper : uncertainWrapperList ) { java . lang . String r = uncertainWrapper . getRuleStr ( ) ; com . dianping . kernel . plugin . SortTool . SortElementWrapper sew = matchRule ( r , allWrapperList ) ; if ( ( sew != null ) && ( duringWrapperList . contains ( sew ) ) ) { if ( r . startsWith ( com . dianping . kernel . plugin . SortTool . beforeOperator ) ) { sew . addBefore ( uncertainWrapper ) ; } else if ( r . startsWith ( com . dianping . kernel . plugin . SortTool . afterOperator ) ) { sew . addAfter ( uncertainWrapper ) ; } else { notMatchWrapperSet . add ( uncertainWrapper ) ; } } else { notMatchWrapperSet . add ( uncertainWrapper ) ; } } java . util . Collections . sort ( beforeWrapperList , new com . dianping . kernel . plugin . SortTool . BeforeOrAfterComparator ( ) ) ; java . util . Collections . sort ( afterWrapperList , new com . dianping . kernel . plugin . SortTool . BeforeOrAfterComparator ( ) ) ; return merge ( beforeWrapperList , duringWrapperList , afterWrapperList , notMatchWrapperSet ) ; } return null ; }
org . junit . Assert . assertArrayEquals ( expectedElementList . toArray ( ) , sel . toArray ( ) )
shouldListAllConstraints ( ) { createUniquenessConstraint ( label1 , propertyKey ) ; createUniquenessConstraint ( label2 , propertyKey ) ; java . util . Set < org . neo4j . internal . kernel . api . schema . constraints . ConstraintDescriptor > constraints = asSet ( storageReader . constraintsGetAll ( ) ) ; java . util . Set < ? > expectedConstraints = asSet ( uniqueConstraintDescriptor ( label1 , propertyKey ) , uniqueConstraintDescriptor ( label2 , propertyKey ) ) ; "<AssertPlaceHolder>" ; } uniqueConstraintDescriptor ( org . neo4j . graphdb . Label , java . lang . String ) { int labelId = labelId ( label ) ; int propKeyId = propertyKeyId ( propertyKey ) ; return org . neo4j . kernel . api . schema . constraints . ConstraintDescriptorFactory . uniqueForLabel ( labelId , propKeyId ) ; }
org . junit . Assert . assertEquals ( expectedConstraints , constraints )
shouldReturnListOfProductsForGivenNameAndId ( ) { sampleProductRepository . save ( new org . springframework . data . elasticsearch . entities . Product ( "1" , "test<sp>product<sp>1" , "How<sp>great<sp>would<sp>it<sp>be<sp>if<sp>we<sp>could<sp>search<sp>for<sp>this<sp>product." , true ) ) ; sampleProductRepository . save ( new org . springframework . data . elasticsearch . entities . Product ( "2" , "test<sp>product<sp>2" , "How<sp>great<sp>would<sp>it<sp>be<sp>if<sp>we<sp>could<sp>search<sp>for<sp>this<sp>other<sp>product." , true ) ) ; java . util . List < org . springframework . data . elasticsearch . entities . Product > products = sampleProductRepository . findByNameAndId ( "product" , "1" ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertThat ( products . size ( ) , org . hamcrest . CoreMatchers . is ( 1 ) )
testAsyncSendEmailCommand2 ( ) { com . ctrip . soa . platform . basesystem . emailservice . v1 . SendEmailResponse response = new com . ctrip . soa . platform . basesystem . emailservice . v1 . SendEmailResponse ( ) ; response . setResultCode ( 1 ) ; when ( client . sendEmail ( any ( ) ) ) . thenReturn ( response ) ; com . ctrip . soa . platform . basesystem . emailservice . v1 . GetEmailStatusResponse getResponse = new com . ctrip . soa . platform . basesystem . emailservice . v1 . GetEmailStatusResponse ( ) ; getResponse . setResultCode ( 1 ) ; when ( client . getEmailStatus ( any ( ) ) ) . thenReturn ( getResponse ) ; com . ctrip . xpipe . service . email . CtripPlatformEmailService . AsyncSendEmailCommand command = new com . ctrip . xpipe . service . email . CtripPlatformEmailService . AsyncSendEmailCommand ( generateEmail ( ) ) ; command . setClient ( client ) ; com . ctrip . xpipe . api . command . CommandFuture future = command . execute ( ) ; "<AssertPlaceHolder>" ; } isSuccess ( ) { return success ; }
org . junit . Assert . assertTrue ( future . isSuccess ( ) )
minTxnHighNoExceptions ( ) { org . apache . hadoop . hive . common . ValidWriteIdList writeIds = new org . apache . hadoop . hive . common . ValidCompactorWriteIdList ( tableName , new long [ 0 ] , new java . util . BitSet ( ) , 5 ) ; org . apache . hadoop . hive . common . ValidWriteIdList . RangeResponse rsp = writeIds . isWriteIdRangeValid ( 7 , 9 ) ; "<AssertPlaceHolder>" ; } isWriteIdRangeValid ( long , long ) { if ( minWriteId > ( highWatermark ) ) { return RangeResponse . NONE ; } else if ( ( ( exceptions . length ) > 0 ) && ( ( exceptions [ 0 ] ) > maxWriteId ) ) { return RangeResponse . ALL ; } long count = java . lang . Math . max ( 0 , ( maxWriteId - ( highWatermark ) ) ) ; for ( long txn : exceptions ) { if ( ( minWriteId <= txn ) && ( txn <= maxWriteId ) ) { count += 1 ; } } if ( count == 0 ) { return RangeResponse . ALL ; } else if ( count == ( ( maxWriteId - minWriteId ) + 1 ) ) { return RangeResponse . NONE ; } else { return RangeResponse . SOME ; } }
org . junit . Assert . assertEquals ( ValidWriteIdList . RangeResponse . NONE , rsp )
testGetTopLevelRecordPositiveIndexOutOfBoundsArrayElement ( ) { org . apache . avro . Schema . Parser parser = new org . apache . avro . Schema . Parser ( ) ; org . apache . avro . Schema schema = parser . parse ( org . pentaho . di . trans . steps . avroinput . AvroInputTest . s_schemaTopLevelRecord ) ; org . apache . avro . io . Decoder decoder ; org . apache . avro . io . DecoderFactory factory = new org . apache . avro . io . DecoderFactory ( ) ; org . apache . avro . generic . GenericData . Record topLevel = new org . apache . avro . generic . GenericData . Record ( schema ) ; org . apache . avro . generic . GenericDatumReader reader = new org . apache . avro . generic . GenericDatumReader ( schema ) ; org . pentaho . di . trans . steps . avroinput . AvroInputMeta . AvroField field = new org . pentaho . di . trans . steps . avroinput . AvroInputMeta . AvroField ( ) ; field . m_fieldName = "test" ; field . m_fieldPath = "$.emails[4]" ; field . m_kettleType = org . pentaho . di . core . row . ValueMeta . getTypeDesc ( ValueMetaInterface . TYPE_STRING ) ; for ( java . lang . String row : org . pentaho . di . trans . steps . avroinput . AvroInputTest . s_jsonDataTopLevelRecord ) { decoder = factory . jsonDecoder ( schema , row ) ; reader . read ( topLevel , decoder ) ; field . init ( 0 ) ; field . reset ( new org . pentaho . di . core . variables . Variables ( ) ) ; java . lang . Object result = field . convertToKettleValue ( topLevel , schema , mock ( org . apache . avro . Schema . class ) , false ) ; "<AssertPlaceHolder>" ; } } convertToKettleValue ( org . apache . avro . generic . GenericData$Record , org . apache . avro . Schema , org . apache . avro . Schema , boolean ) { if ( record == null ) { return null ; } if ( ( m_tempParts . size ( ) ) == 0 ) { throw new org . pentaho . di . core . exception . KettleException ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.MalformedPathRecord" ) ) ; } java . lang . String part = m_tempParts . remove ( 0 ) ; if ( ( part . charAt ( 0 ) ) == '[' ) { throw new org . pentaho . di . core . exception . KettleException ( ( ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.InvalidPath" ) ) + ( m_tempParts ) ) ) ; } if ( ( part . indexOf ( '[' ) ) > 0 ) { java . lang . String arrayPart = part . substring ( part . indexOf ( '[' ) ) ; part = part . substring ( 0 , part . indexOf ( '[' ) ) ; m_tempParts . add ( 0 , arrayPart ) ; } org . apache . avro . Schema . Field fieldS = s . getField ( part ) ; if ( ( fieldS == null ) && ( ! ignoreMissing ) ) { throw new org . pentaho . di . core . exception . KettleException ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.NonExistentField" , part ) ) ; } java . lang . Object field = record . get ( part ) ; if ( field == null ) { fieldS = defaultSchema . getField ( part ) ; if ( ( fieldS == null ) || ( ( fieldS . defaultValue ( ) ) == null ) ) { return null ; } field = fieldS . defaultValue ( ) ; } org . apache . avro . Schema . Type fieldT = fieldS . schema ( ) . getType ( ) ; org . apache . avro . Schema fieldSchema = fieldS . schema ( ) ; if ( fieldT == ( Schema . Type . UNION ) ) { if ( field instanceof org . apache . avro . generic . GenericContainer ) { fieldSchema = ( ( org . apache . avro . generic . GenericContainer ) ( field ) ) . getSchema ( ) ; fieldT = fieldSchema . getType ( ) ; } else { if ( field instanceof java . util . Map ) { org . apache . avro . Schema mapSchema = null ; for ( org . apache . avro . Schema ts : fieldSchema . getTypes ( ) ) { if ( ( ts . getType ( ) ) == ( Schema . Type . MAP ) ) { mapSchema = ts ; break ; } } if ( mapSchema == null ) { throw new org . pentaho . di . core . exception . KettleException ( org . pentaho . di . i18n . BaseMessages . getString ( org . pentaho . di . trans . steps . avroinput . AvroInputMeta . PKG , "AvroInput.Error.UnableToFindSchemaForUnionMap" ) ) ; } fieldSchema = mapSchema ; fieldT = Schema . Type . MAP ; } else { if ( ( m_tempValueMeta . getType ( ) ) != ( org . pentaho . di . core . row . ValueMetaInterface . TYPE_STRING ) ) { fieldSchema = org . pentaho . di . trans . steps . avroinput . AvroInputData . checkUnion ( fieldSchema ) ; fieldT = fieldSchema
org . junit . Assert . assertTrue ( ( result == null ) )
testBuildWithParametersWithOrderBy ( ) { org . lnu . is . domain . department . Department department = new org . lnu . is . domain . department . Department ( ) ; org . lnu . is . domain . specialty . Specialty specialty = new org . lnu . is . domain . specialty . Specialty ( ) ; org . lnu . is . domain . department . specialty . DepartmentSpecialty context = new org . lnu . is . domain . department . specialty . DepartmentSpecialty ( ) ; context . setDepartment ( department ) ; context . setSpecialty ( specialty ) ; org . lnu . is . pagination . OrderBy orderBy1 = new org . lnu . is . pagination . OrderBy ( "department" , org . lnu . is . pagination . OrderByType . ASC ) ; org . lnu . is . pagination . OrderBy orderBy2 = new org . lnu . is . pagination . OrderBy ( "specialty" , org . lnu . is . pagination . OrderByType . DESC ) ; java . util . List < org . lnu . is . pagination . OrderBy > orders = java . util . Arrays . asList ( orderBy1 , orderBy2 ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>DepartmentSpecialty<sp>e<sp>WHERE<sp>(<sp>e.department<sp>=<sp>:department<sp>AND<sp>e.specialty<sp>=<sp>:specialty<sp>)<sp>AND<sp>e.status=:status<sp>AND<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>ORDER<sp>BY<sp>e.department<sp>ASC,<sp>e.specialty<sp>DESC" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . department . specialty . DepartmentSpecialty > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; pagedSearch . setOrders ( orders ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setOrders ( java . util . List ) { this . orders = orders ; }
org . junit . Assert . assertEquals ( expectedQuery , actualQuery )
deveObterCampoeValorNotaComoFoiSetado ( ) { final com . fincatto . documentofiscal . nfe400 . classes . nota . NFNotaInfoItemProduto produto = new com . fincatto . documentofiscal . nfe400 . classes . nota . NFNotaInfoItemProduto ( ) ; final com . fincatto . documentofiscal . nfe400 . classes . NFProdutoCompoeValorNota compoeValorNota = com . fincatto . documentofiscal . nfe400 . classes . NFProdutoCompoeValorNota . SIM ; produto . setCampoeValorNota ( compoeValorNota ) ; "<AssertPlaceHolder>" ; } getCompoeValorNota ( ) { return this . compoeValorNota ; }
org . junit . Assert . assertEquals ( compoeValorNota , produto . getCompoeValorNota ( ) )
getAllPatientIdentifierTypes_shouldOrderAsDefaultComparator ( ) { java . util . List < org . openmrs . PatientIdentifierType > list = org . openmrs . api . PatientServiceTest . patientService . getAllPatientIdentifierTypes ( ) ; java . util . List < org . openmrs . PatientIdentifierType > sortedList = new java . util . ArrayList ( list ) ; sortedList . sort ( new org . openmrs . comparator . PatientIdentifierTypeDefaultComparator ( ) ) ; "<AssertPlaceHolder>" ; } getAllPatientIdentifierTypes ( ) { return org . openmrs . api . context . Context . getPatientService ( ) . getAllPatientIdentifierTypes ( false ) ; }
org . junit . Assert . assertEquals ( sortedList , list )
getIndexedKey ( ) { java . lang . Runnable runnable = new java . lang . Runnable ( ) { @ com . cetsoft . imcache . cache . search . SuppressWarnings ( "unused" ) java . lang . String runnable = "runnable" ; public void run ( ) { } } ; "<AssertPlaceHolder>" ; } getIndexedKey ( java . lang . String , V ) { try { java . lang . reflect . Field field = value . getClass ( ) . getDeclaredField ( attributeName ) ; field . setAccessible ( true ) ; return field . get ( value ) ; } catch ( java . lang . Exception e ) { throw new com . cetsoft . imcache . cache . search . AttributeException ( e ) ; } }
org . junit . Assert . assertEquals ( "runnable" , handler . getIndexedKey ( "runnable" , runnable ) )
serializeNull ( ) { org . eclipse . jgit . lib . ObjectId deserialized = writeAndReadBackFromTempFile ( null ) ; "<AssertPlaceHolder>" ; } writeAndReadBackFromTempFile ( org . eclipse . jgit . lib . ObjectId ) { java . io . File file = java . io . File . createTempFile ( "ObjectIdSerializerTest_" , "" ) ; try ( java . io . OutputStream out = new java . io . FileOutputStream ( file ) ) { if ( objectId == null ) { org . eclipse . jgit . lib . ObjectIdSerializer . write ( out , objectId ) ; } else { org . eclipse . jgit . lib . ObjectIdSerializer . writeWithoutMarker ( out , objectId ) ; } } try ( java . io . InputStream in = new java . io . FileInputStream ( file ) ) { if ( objectId == null ) { return org . eclipse . jgit . lib . ObjectIdSerializer . read ( in ) ; } else { return org . eclipse . jgit . lib . ObjectIdSerializer . readWithoutMarker ( in ) ; } } }
org . junit . Assert . assertNull ( deserialized )
testStdOutNotPrintedIfNotShouldRecordStdoutEvenIfVerbose ( ) { com . facebook . buck . shell . ShellStep command = com . facebook . buck . shell . ShellStepTest . createCommand ( false , false ) ; com . facebook . buck . util . ProcessExecutorParams params = com . facebook . buck . shell . ShellStepTest . createParams ( ) ; com . facebook . buck . util . FakeProcess process = new com . facebook . buck . util . FakeProcess ( com . facebook . buck . shell . ShellStepTest . EXIT_SUCCESS , com . facebook . buck . shell . ShellStepTest . OUTPUT_MSG , com . facebook . buck . shell . ShellStepTest . ERROR_MSG ) ; com . facebook . buck . testutil . TestConsole console = new com . facebook . buck . testutil . TestConsole ( com . facebook . buck . util . Verbosity . ALL ) ; com . facebook . buck . core . build . execution . context . ExecutionContext context = com . facebook . buck . shell . ShellStepTest . createContext ( com . google . common . collect . ImmutableMap . of ( params , process ) , console ) ; command . launchAndInteractWithProcess ( context , params ) ; "<AssertPlaceHolder>" ; } getTextWrittenToStdErr ( ) { com . facebook . buck . util . CapturingPrintStream stream = ( ( com . facebook . buck . util . CapturingPrintStream ) ( getStdErr ( ) . getRawStream ( ) ) ) ; return stream . getContentsAsString ( Charsets . UTF_8 ) ; }
org . junit . Assert . assertEquals ( "" , console . getTextWrittenToStdErr ( ) )
testCollectionRequiredInRole ( ) { append ( "[groups]" , globalPolicyFile ) ; append ( "group<sp>=<sp>some_role" , globalPolicyFile ) ; append ( "[roles]" , globalPolicyFile ) ; append ( "some_role<sp>=<sp>action=query" , globalPolicyFile ) ; org . apache . sentry . policy . common . PolicyEngine policy = new org . apache . sentry . policy . search . SearchPolicyFileBackend ( globalPolicyFile . getPath ( ) ) ; com . google . common . collect . ImmutableSet < java . lang . String > permissions = policy . getPrivileges ( com . google . common . collect . Sets . newHashSet ( "group" ) , ActiveRoleSet . ALL ) ; "<AssertPlaceHolder>" ; } toString ( ) { return PolicyConstants . AUTHORIZABLE_JOINER . join ( parts ) ; }
org . junit . Assert . assertTrue ( permissions . toString ( ) , permissions . isEmpty ( ) )
headingAndParagraphs ( ) { java . lang . StringBuilder stb = new java . lang . StringBuilder ( ) ; stb . append ( "<H<sp>LVL=\"1\"><STEXT><PDAT>HEADING<sp>TEXT</PDAT></STEXT></H>\n" ) ; stb . append ( "<PARA<sp>ID=\"P-00003\"<sp>LVL=\"0\"><PTEXT><PDAT>pargraph<sp>text.</PDAT></PTEXT></PARA>" ) ; java . lang . String input = stb . toString ( ) ; java . lang . StringBuilder expectStb = new java . lang . StringBuilder ( ) ; expectStb . append ( "<h2<sp>level=\"1\">HEADING<sp>TEXT</h2>\n" ) ; expectStb . append ( "<p<sp>id=\"P-00003\"<sp>level=\"0\">pargraph<sp>text.</p>" ) ; java . lang . String expect = expectStb . toString ( ) ; java . lang . String actual = format . getSimpleHtml ( input ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( expect , actual )
testRadicalSiteRrBetaReaction ( ) { org . openscience . cdk . reaction . IReactionProcess type = new org . openscience . cdk . reaction . type . RadicalSiteRrBetaReaction ( ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( type )
testContradiction ( ) { org . infinispan . query . dsl . QueryFactory qf = org . infinispan . all . embeddedquery . QueryDslConditionsTest . getQueryFactory ( ) ; org . infinispan . query . dsl . Query q = qf . from ( org . infinispan . all . embeddedquery . QueryDslConditionsTest . getModelFactory ( ) . getUserImplClass ( ) ) . having ( "name" ) . gt ( "A" ) . and ( ) . having ( "name" ) . lte ( "A" ) . build ( ) ; java . util . List < org . infinispan . all . embeddedquery . testdomain . User > list = q . list ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return counterListener . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( list . isEmpty ( ) )
testNoRootAttachChange ( ) { java . util . List < com . vaadin . flow . internal . change . NodeChange > changes = collectChangesExceptChildrenAddRemove ( ) ; "<AssertPlaceHolder>" ; } collectChangesExceptChildrenAddRemove ( ) { java . util . List < com . vaadin . flow . internal . change . NodeChange > changes = new java . util . ArrayList ( ) ; tree . collectChanges ( ( change ) -> { if ( ( ( change instanceof com . vaadin . flow . internal . change . ListAddChange ) || ( change instanceof com . vaadin . flow . internal . change . ListRemoveChange ) ) && ( ( ( ( com . vaadin . flow . internal . change . NodeFeatureChange ) ( change ) ) . getFeature ( ) ) == ( . class ) ) ) { return ; } else { changes . add ( change ) ; } } ) ; return changes ; }
org . junit . Assert . assertEquals ( java . util . Collections . emptyList ( ) , changes )
testPropertyDeltaBatching ( ) { int n = 10 ; int base = 100 ; for ( int i = 0 ; i < n ; i ++ ) { org . nuxeo . ecm . core . api . DocumentModel doc = session . createDocumentModel ( "/" , ( "doc" + i ) , "MyDocType" ) ; doc . setPropertyValue ( "my:integer" , java . lang . Long . valueOf ( base ) ) ; doc = session . createDocument ( doc ) ; } session . save ( ) ; for ( int i = 0 ; i < n ; i ++ ) { org . nuxeo . ecm . core . api . DocumentModel doc = session . getDocument ( new org . nuxeo . ecm . core . api . PathRef ( ( "/doc" + i ) ) ) ; java . io . Serializable value ; if ( i < ( n / 2 ) ) { value = java . lang . Long . valueOf ( i ) ; } else { value = org . nuxeo . ecm . core . api . model . DeltaLong . valueOf ( java . lang . Long . valueOf ( base ) , i ) ; } doc . setPropertyValue ( "my:integer" , value ) ; if ( ( i % 2 ) == 0 ) { doc . setPropertyValue ( "my:string" , ( "foo" + i ) ) ; } doc = session . saveDocument ( doc ) ; } session . save ( ) ; reopenSession ( ) ; for ( int i = 0 ; i < n ; i ++ ) { org . nuxeo . ecm . core . api . DocumentModel doc = session . getDocument ( new org . nuxeo . ecm . core . api . PathRef ( ( "/doc" + i ) ) ) ; java . io . Serializable value = doc . getPropertyValue ( "my:integer" ) ; java . io . Serializable expected ; if ( i < ( n / 2 ) ) { expected = java . lang . Long . valueOf ( i ) ; } else { expected = java . lang . Long . valueOf ( ( base + i ) ) ; } "<AssertPlaceHolder>" ; } } getName ( ) { return name ; }
org . junit . Assert . assertEquals ( doc . getName ( ) , expected , value )
menuShouldHaveItems ( ) { final java . util . List < lcmc . common . ui . utils . UpdatableItem > items = resourceMenu . getPulldownMenu ( resourceInfoStub ) ; verify ( menuStub , never ( ) ) . enablePredicate ( ( ( lcmc . common . domain . EnablePredicate ) ( anyObject ( ) ) ) ) ; verify ( menuStub , times ( 2 ) ) . onUpdate ( ( ( java . lang . Runnable ) ( anyObject ( ) ) ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return hosts . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , items . size ( ) )
testCreateSettingsPanel ( ) { System . out . println ( "createSettingsPanel" ) ; kg . apc . jmeter . vizualizers . PerfMonGui instance = new kg . apc . jmeter . vizualizers . PerfMonGui ( ) ; kg . apc . jmeter . vizualizers . JSettingsPanel result = instance . createSettingsPanel ( ) ; "<AssertPlaceHolder>" ; } createSettingsPanel ( ) { return new kg . apc . jmeter . vizualizers . JSettingsPanel ( this , ( ( ( ( ( ( ( JSettingsPanel . TIMELINE_OPTION ) | ( JSettingsPanel . GRADIENT_OPTION ) ) | ( JSettingsPanel . FINAL_ZEROING_OPTION ) ) | ( JSettingsPanel . LIMIT_POINT_OPTION ) ) | ( JSettingsPanel . MAXY_OPTION ) ) | ( JSettingsPanel . RELATIVE_TIME_OPTION ) ) | ( JSettingsPanel . MARKERS_OPTION ) ) ) ; }
org . junit . Assert . assertNotNull ( result )
testPasswordDifferentKeys ( ) { ch . bfh . unicrypt . crypto . schemes . encryption . classes . AESEncryptionScheme aes = ch . bfh . unicrypt . crypto . schemes . encryption . classes . AESEncryptionScheme . getInstance ( AESEncryptionScheme . KeyLength . KEY128 , AESEncryptionScheme . Mode . ECB , ch . bfh . unicrypt . helper . array . classes . ByteArray . getInstance ( "00|00|00|00|00|00|00|00|00|00|00|00|00|00|00|00" ) ) ; ch . bfh . unicrypt . math . algebra . general . interfaces . Element < ch . bfh . unicrypt . helper . array . classes . ByteArray > key = aes . getSecretKeyGenerator ( ) . generateSecretKey ( "This<sp>is<sp>the<sp>Test" ) ; ch . bfh . unicrypt . crypto . schemes . encryption . classes . AESEncryptionScheme aes2 = ch . bfh . unicrypt . crypto . schemes . encryption . classes . AESEncryptionScheme . getInstance ( AESEncryptionScheme . KeyLength . KEY128 , AESEncryptionScheme . Mode . ECB , ch . bfh . unicrypt . helper . array . classes . ByteArray . getInstance ( "00|00|00|00|00|00|00|00|00|00|00|00|00|00|00|00" ) ) ; ch . bfh . unicrypt . math . algebra . general . interfaces . Element < ch . bfh . unicrypt . helper . array . classes . ByteArray > key2 = aes2 . getSecretKeyGenerator ( ) . generateSecretKey ( "Thas<sp>is<sp>the<sp>Test" ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( obj == null ) { return false ; } if ( obj instanceof ch . bfh . unicrypt . helper . prime . Factorization ) { final ch . bfh . unicrypt . helper . prime . Factorization other = ( ( ch . bfh . unicrypt . helper . prime . Factorization ) ( obj ) ) ; return this . value . equals ( other . value ) ; } return false ; }
org . junit . Assert . assertFalse ( key . equals ( key2 ) )
find_with_show_only_running_shows_if_not_enough_supports_and_if_time_has_not_ended_inclusive ( ) { testHelper . create ( new fi . om . initiative . dao . TestHelper . InitiativeDraft ( userId ) . withState ( InitiativeState . ACCEPTED ) . withSupportCount ( 49 ) . isRunning ( fi . om . initiative . dao . InitiativeDaoSearchTest . day_before_yesterday , fi . om . initiative . dao . InitiativeDaoSearchTest . far_in_the_future ) ) ; testHelper . create ( new fi . om . initiative . dao . TestHelper . InitiativeDraft ( userId ) . withState ( InitiativeState . ACCEPTED ) . withSupportCount ( 50 ) . isRunning ( fi . om . initiative . dao . InitiativeDaoSearchTest . day_before_yesterday , fi . om . initiative . dao . InitiativeDaoSearchTest . far_in_the_future ) ) ; fi . om . initiative . dto . search . InitiativeSearch search = initiativeSearch ( ) ; search . setShow ( Show . running ) ; fi . om . initiative . dto . InitiativeSettings . MinSupportCountSettings supportCountSettings = new fi . om . initiative . dto . InitiativeSettings . MinSupportCountSettings ( 50 , org . joda . time . Days . days ( 2 ) ) ; java . util . List < fi . om . initiative . dto . initiative . InitiativeInfo > result = initiativeDao . findInitiatives ( search , userId , supportCountSettings ) . list ; "<AssertPlaceHolder>" ; } size ( ) { return hashMap . size ( ) ; }
org . junit . Assert . assertThat ( result . size ( ) , org . hamcrest . CoreMatchers . is ( 2 ) )
serviceUnregistrationClusterSingletonServiceGroupTwoServicesTest ( ) { singletonServiceGroup . initialize ( ) ; verify ( mockEosService ) . registerCandidate ( org . opendaylight . mdsal . singleton . dom . impl . ClusterSingletonServiceGroupImplTest . MAIN_ENTITY ) ; singletonServiceGroup . registerService ( firstReg ) ; singletonServiceGroup . registerService ( secondReg ) ; "<AssertPlaceHolder>" ; verify ( mockClusterSingletonService , never ( ) ) . closeServiceInstance ( ) ; } unregisterService ( org . opendaylight . mdsal . singleton . common . api . ClusterSingletonServiceRegistration ) { verifyNoSuccessor ( ) ; services . remove ( reg ) ; org . opendaylight . mdsal . singleton . dom . impl . PlaceholderGroup . LOG . debug ( "{}:<sp>removed<sp>service<sp>{}" , this , reg . getInstance ( ) ) ; return null ; }
org . junit . Assert . assertNull ( singletonServiceGroup . unregisterService ( firstReg ) )
shouldFindProjectForCustomersOnlyActive ( ) { java . util . ArrayList < net . rrm . ehour . persistence . project . dao . Customer > ids = new java . util . ArrayList ( java . util . Arrays . asList ( new net . rrm . ehour . persistence . project . dao . Customer ( 30 ) , new net . rrm . ehour . persistence . project . dao . Customer ( 10 ) ) ) ; java . util . List < net . rrm . ehour . persistence . project . dao . Project > r = projectDAO . findProjectForCustomers ( ids , true ) ; "<AssertPlaceHolder>" ; } size ( ) { return joinTables . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , r . size ( ) )
testInvokeCallbackWithTypedArray_Uint8ClampledArray ( ) { com . eclipsesource . v8 . JavaCallback callback = new com . eclipsesource . v8 . JavaCallback ( ) { @ com . eclipsesource . v8 . Override public com . eclipsesource . v8 . Boolean invoke ( final com . eclipsesource . v8 . V8Object receiver , final com . eclipsesource . v8 . V8Array parameters ) { com . eclipsesource . v8 . V8TypedArray result = ( ( com . eclipsesource . v8 . utils . TypedArray ) ( com . eclipsesource . v8 . utils . V8ObjectUtils . getValue ( parameters , 0 ) ) ) . getV8TypedArray ( ) ; try { return ( result . getType ( ) ) == ( V8Value . UNSIGNED_INT_8_CLAMPED_ARRAY ) ; } finally { result . close ( ) ; } } } ; v8 . registerJavaMethod ( callback , "callback" ) ; java . lang . Object result = v8 . executeScript ( "callback(new<sp>Uint8ClampedArray(24));" ) ; "<AssertPlaceHolder>" ; } executeScript ( java . lang . String ) { return executeScript ( script , null , 0 ) ; }
org . junit . Assert . assertTrue ( ( ( java . lang . Boolean ) ( result ) ) )
payBillUsingVisa ( ) { com . captaindebug . bridge . ShoppingCart instance = new com . captaindebug . bridge . ShoppingCart ( ) ; com . captaindebug . telldontask . Item a = new com . captaindebug . telldontask . Item ( "gloves" , 23.43 ) ; instance . addItem ( a ) ; com . captaindebug . telldontask . Item b = new com . captaindebug . telldontask . Item ( "hat" , 10.99 ) ; instance . addItem ( b ) ; java . util . Date expiryDate = getCardExpireyDate ( ) ; com . captaindebug . payment . PaymentMethod visa = new com . captaindebug . payment . Visa ( "CaptainDebug" , "1234234534564567" , expiryDate ) ; boolean result = instance . pay ( visa ) ; "<AssertPlaceHolder>" ; } pay ( double ) { return true ; }
org . junit . Assert . assertTrue ( result )
testMaxFunctionQuantityType ( ) { items . add ( createNumberItem ( "TestItem1" , javax . measure . quantity . Temperature . class , new org . eclipse . smarthome . core . library . types . QuantityType < javax . measure . quantity . Temperature > ( "100<sp>C" ) ) ) ; items . add ( createNumberItem ( "TestItem2" , javax . measure . quantity . Temperature . class , UnDefType . NULL ) ) ; items . add ( createNumberItem ( "TestItem3" , javax . measure . quantity . Temperature . class , new org . eclipse . smarthome . core . library . types . QuantityType < javax . measure . quantity . Temperature > ( "200<sp>C" ) ) ) ; items . add ( createNumberItem ( "TestItem4" , javax . measure . quantity . Temperature . class , UnDefType . UNDEF ) ) ; items . add ( createNumberItem ( "TestItem5" , javax . measure . quantity . Temperature . class , new org . eclipse . smarthome . core . library . types . QuantityType < javax . measure . quantity . Temperature > ( "300<sp>C" ) ) ) ; function = new org . eclipse . smarthome . core . library . types . QuantityTypeArithmeticGroupFunction . Max ( javax . measure . quantity . Temperature . class ) ; org . eclipse . smarthome . core . types . State state = function . calculate ( items ) ; "<AssertPlaceHolder>" ; } createNumberItem ( java . lang . String , java . lang . Class , org . eclipse . smarthome . core . types . State ) { org . eclipse . smarthome . core . library . items . NumberItem item = new org . eclipse . smarthome . core . library . items . NumberItem ( ( ( ( org . eclipse . smarthome . core . library . CoreItemFactory . NUMBER ) + ":" ) + ( dimension . getSimpleName ( ) ) ) , name ) ; item . setUnitProvider ( unitProvider ) ; item . setState ( state ) ; return item ; }
org . junit . Assert . assertEquals ( new org . eclipse . smarthome . core . library . types . QuantityType < javax . measure . quantity . Temperature > ( "300<sp>C" ) , state )
whenAddingTwoVehiclesWithSameTypeIdAndLocation_getAvailableVehicleShouldReturnOnlyOneOfThem ( ) { jsprit . core . problem . vehicle . VehicleTypeImpl type = VehicleTypeImpl . Builder . newInstance ( "standard" ) . build ( ) ; jsprit . core . problem . vehicle . Vehicle v1 = VehicleImpl . Builder . newInstance ( "v1" ) . setStartLocation ( jsprit . core . problem . Location . newInstance ( "loc" ) ) . setType ( type ) . build ( ) ; jsprit . core . problem . vehicle . Vehicle v2 = VehicleImpl . Builder . newInstance ( "v2" ) . setStartLocation ( jsprit . core . problem . Location . newInstance ( "loc" ) ) . setType ( type ) . build ( ) ; jsprit . core . problem . vehicle . VehicleFleetManager fleetManager = new jsprit . core . problem . vehicle . FiniteFleetManagerFactory ( java . util . Arrays . asList ( v1 , v2 ) ) . createFleetManager ( ) ; java . util . Collection < jsprit . core . problem . vehicle . Vehicle > vehicles = fleetManager . getAvailableVehicles ( ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( 1 , vehicles . size ( ) )
testGetChildFoldersNull ( ) { net . holmes . core . business . version . VersionManager versionManager = createMock ( net . holmes . core . business . version . VersionManager . class ) ; replay ( versionManager ) ; net . holmes . core . backend . handler . UtilHandler utilHandler = new net . holmes . core . backend . handler . UtilHandler ( versionManager ) ; java . util . Collection < net . holmes . core . backend . handler . UtilHandler . Folder > folders = utilHandler . getChildFolders ( null ) ; "<AssertPlaceHolder>" ; verify ( versionManager ) ; } getChildFolders ( java . lang . String ) { java . util . Collection < net . holmes . core . backend . handler . UtilHandler . Folder > folders = new java . util . ArrayList ( ) ; if ( ( parentPath == null ) || ( "none" . equals ( parentPath ) ) ) { java . io . File userHomeDir = new java . io . File ( net . holmes . core . backend . handler . USER_HOME . getValue ( ) ) ; folders . add ( new net . holmes . core . backend . handler . UtilHandler . Folder ( userHomeDir . getName ( ) , userHomeDir . getAbsolutePath ( ) ) ) ; java . nio . file . FileSystems . getDefault ( ) . getRootDirectories ( ) . forEach ( ( root ) -> folders . add ( new net . holmes . core . backend . handler . Folder ( root . toString ( ) , root . toString ( ) ) ) ) ; } else { folders . addAll ( listChildren ( parentPath , false ) . stream ( ) . map ( ( child ) -> new net . holmes . core . backend . handler . Folder ( child . getName ( ) , child . getAbsolutePath ( ) ) ) . collect ( toList ( ) ) ) ; } return folders ; }
org . junit . Assert . assertNotNull ( folders )
removeTimestampFromId ( ) { java . lang . String productId = PRODUCT_ID ; java . lang . String result = org . oscm . converter . TimeStampUtil . removeTimestampFromId ( productId ) ; "<AssertPlaceHolder>" ; } removeTimestampFromId ( java . lang . String ) { if ( ( id != null ) && ( id . contains ( "#" ) ) ) { return id . substring ( 0 , id . lastIndexOf ( "#" ) ) ; } return id ; }
org . junit . Assert . assertEquals ( PRODUCT_ID , result )
ClaimRef ( ) { java . lang . String input = "<claim-ref<sp>idref=\"CLM-00001\">claim<sp>1</claim-ref>" ; java . lang . String expect = "<a<sp>idref=\"CLM-00001\"<sp>id=\"CR-0001\"<sp>class=\"claim\">claim<sp>1</a>" ; java . lang . String actual = format . getSimpleHtml ( input ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( expect , actual )
testFetchNextRecordReaderThrowsException ( ) { org . apache . flink . api . java . hadoop . mapreduce . HadoopInputFormatTest . DummyRecordReader recordReader = mock ( org . apache . flink . api . java . hadoop . mapreduce . HadoopInputFormatTest . DummyRecordReader . class ) ; when ( recordReader . nextKeyValue ( ) ) . thenThrow ( new java . lang . InterruptedException ( ) ) ; org . apache . flink . api . java . hadoop . mapreduce . HadoopInputFormat < java . lang . String , java . lang . Long > hadoopInputFormat = setupHadoopInputFormat ( new org . apache . flink . api . java . hadoop . mapreduce . HadoopInputFormatTest . DummyInputFormat ( ) , org . apache . hadoop . mapreduce . Job . getInstance ( ) , recordReader ) ; exception . expect ( java . io . IOException . class ) ; hadoopInputFormat . fetchNext ( ) ; "<AssertPlaceHolder>" ; } fetchNext ( ) { try { this . hasNext = this . recordReader . nextKeyValue ( ) ; } catch ( java . lang . InterruptedException e ) { throw new java . io . IOException ( "Could<sp>not<sp>fetch<sp>next<sp>KeyValue<sp>pair." , e ) ; } finally { this . fetched = true ; } }
org . junit . Assert . assertThat ( hadoopInputFormat . hasNext , org . hamcrest . core . Is . is ( true ) )
test59getSecureServicePoliciesIfUpdatedSuccess ( ) { javax . servlet . http . HttpServletRequest request = org . mockito . Mockito . mock ( javax . servlet . http . HttpServletRequest . class ) ; java . lang . Long lastKnownVersion = 1L ; java . lang . String pluginId = "1" ; org . apache . ranger . entity . XXService xService = xService ( ) ; org . apache . ranger . entity . XXServiceDef xServiceDef = serviceDef ( ) ; xServiceDef . setImplclassname ( "org.apache.ranger.services.kms.RangerServiceKMS" ) ; java . lang . String serviceName = xService . getName ( ) ; org . apache . ranger . plugin . model . RangerService rs = rangerService ( ) ; org . apache . ranger . plugin . util . ServicePolicies sp = servicePolicies ( ) ; org . apache . ranger . db . XXServiceDefDao xServiceDefDao = org . mockito . Mockito . mock ( org . apache . ranger . db . XXServiceDefDao . class ) ; org . mockito . Mockito . when ( serviceUtil . isValidService ( serviceName , request ) ) . thenReturn ( true ) ; org . mockito . Mockito . when ( daoManager . getXXService ( ) ) . thenReturn ( xServiceDao ) ; org . mockito . Mockito . when ( xServiceDao . findByName ( serviceName ) ) . thenReturn ( xService ) ; org . mockito . Mockito . when ( daoManager . getXXServiceDef ( ) ) . thenReturn ( xServiceDefDao ) ; org . mockito . Mockito . when ( xServiceDefDao . getById ( xService . getType ( ) ) ) . thenReturn ( xServiceDef ) ; org . mockito . Mockito . when ( svcStore . getServiceByNameForDP ( serviceName ) ) . thenReturn ( rs ) ; org . mockito . Mockito . when ( bizUtil . isUserAllowed ( rs , ServiceREST . Allowed_User_List_For_Grant_Revoke ) ) . thenReturn ( true ) ; org . mockito . Mockito . when ( svcStore . getServicePoliciesIfUpdated ( org . mockito . Mockito . anyString ( ) , org . mockito . Mockito . anyLong ( ) , org . mockito . Mockito . anyBoolean ( ) ) ) . thenReturn ( sp ) ; org . mockito . Mockito . when ( zoneStore . getSecurityZonesForService ( serviceName ) ) . thenReturn ( null ) ; org . apache . ranger . plugin . util . ServicePolicies dbServiceSecurePolicies = serviceREST . getSecureServicePoliciesIfUpdated ( serviceName , lastKnownVersion , 0L , pluginId , "" , "" , true , request ) ; "<AssertPlaceHolder>" ; org . mockito . Mockito . verify ( serviceUtil ) . isValidService ( serviceName , request ) ; org . mockito . Mockito . verify ( xServiceDao ) . findByName ( serviceName ) ; org . mockito . Mockito . verify ( xServiceDefDao ) . getById ( xService . getType ( ) ) ; org . mockito . Mockito . verify ( svcStore ) . getServiceByNameForDP ( serviceName ) ; org . mockito . Mockito . verify ( bizUtil ) . isUserAllowed ( rs , ServiceREST . Allowed_User_List_For_Grant_Revoke ) ; org . mockito . Mockito . verify ( svcStore ) . getServicePoliciesIfUpdated ( serviceName , lastKnownVersion , false ) ; } getSecureServicePoliciesIfUpdated ( java . lang . String , java . lang . Long , java . lang . Long , java . lang . String , java . lang . String , java . lang . String , java . lang . Boolean , javax . servlet . http . HttpServletRequest ) { if ( org . apache . ranger . rest . ServiceREST . LOG . isDebugEnabled ( ) ) { org . apache . ranger . rest . ServiceREST . LOG . debug ( ( ( ( ( ( ( ( ( ( ( ( ( "secure" 5 + serviceName ) + ",<sp>" ) + lastKnownVersion ) + ",<sp>" ) + lastActivationTime ) + ",<sp>" ) + pluginId ) + ",<sp>" ) + clusterName ) + ",<sp>" ) + supportsPolicyDeltas ) + "secure" 1 ) ) ; } org . apache . ranger . plugin . util . ServicePolicies ret = null ; int httpCode = javax . servlet . http . HttpServletResponse . SC_OK ; java . lang . String logMsg = null ; org . apache . ranger . plugin . util . RangerPerfTracer perf = null ; boolean isAllowed = false ; boolean isAdmin = bizUtil . isAdmin ( ) ; boolean isKeyAdmin = bizUtil . isKeyAdmin ( ) ; request . setAttribute ( "downloadPolicy" , "secure" ) ; java . lang . Long downloadedVersion = null ; boolean isValid = false ; try { isValid = serviceUtil . isValidService ( serviceName , request ) ; } catch ( javax . ws . rs . WebApplicationException webException ) { httpCode = webException . getResponse ( ) . getStatus ( ) ; logMsg = webException . getResponse ( ) . getEntity ( ) . toString ( ) ; } catch ( java . lang . Exception e ) { httpCode = javax . servlet . http . HttpServletResponse . SC_BAD_REQUEST ; logMsg = e . getMessage ( ) ; } if ( isValid ) { if ( lastKnownVersion == null ) { lastKnownVersion = java . lang . Long . valueOf ( ( - 1 ) ) ; } try { if ( org . apache . ranger . plugin . util . RangerPerfTracer . isPerfTraceEnabled ( org . apache . ranger . rest . ServiceREST . PERF_LOG ) ) { perf = org . apache . ranger . plugin . util . RangerPerfTracer . getPerfTracer ( org . apache . ranger . rest . ServiceREST . PERF_LOG , ( ( ( ( ( ( "ServiceREST.getSecureServicePoliciesIfUpdated(serviceName=" + serviceName ) + ",lastKnownVersion=" ) + lastKnownVersion ) + ",lastActivationTime=" ) + lastActivationTime ) + "secure" 1 ) ) ; } org . apache . ranger . entity . XXService xService = daoManager . getXXService ( ) . findByName ( serviceName
org . junit . Assert . assertNotNull ( dbServiceSecurePolicies )
listcompactions ( ) { final java . lang . String table = name . getMethodName ( ) ; ts . exec ( ( "createtable<sp>" + table ) , true ) ; ts . exec ( ( ( "insert<sp>a<sp>cf<sp>cq<sp>value" 2 + table ) + "<sp>-s<sp>table.iterator.minc.slow=30,org.apache.accumulo.test.functional.SlowIterator" ) , true ) ; ts . exec ( ( ( "insert<sp>a<sp>cf<sp>cq<sp>value" 2 + table ) + "<sp>-s<sp>table.iterator.minc.slow.opt.sleepTime=1000" ) , true ) ; ts . exec ( "insert<sp>a<sp>cf<sp>cq<sp>value" , true ) ; ts . exec ( "insert<sp>b<sp>cf<sp>cq<sp>value" , true ) ; ts . exec ( "insert<sp>c<sp>cf<sp>cq<sp>value" , true ) ; ts . exec ( "insert<sp>a<sp>cf<sp>cq<sp>value" 3 , true ) ; ts . exec ( ( "flush<sp>-t<sp>" + table ) , true ) ; ts . exec ( "sleep<sp>0.2" , true ) ; ts . exec ( "insert<sp>a<sp>cf<sp>cq<sp>value" 0 , true , "default_tablet" ) ; java . lang . String [ ] lines = ts . output . get ( ) . split ( "insert<sp>a<sp>cf<sp>cq<sp>value" 4 ) ; java . lang . String last = lines [ ( ( lines . length ) - 1 ) ] ; java . lang . String [ ] parts = last . split ( "\\|" ) ; "<AssertPlaceHolder>" ; ts . exec ( ( "insert<sp>a<sp>cf<sp>cq<sp>value" 1 + table ) , true ) ; } split ( byte [ ] ) { if ( ( ( sp != null ) && ( ( extent . getEndRow ( ) ) != null ) ) && ( extent . getEndRow ( ) . equals ( new org . apache . hadoop . io . Text ( sp ) ) ) ) { throw new java . lang . IllegalArgumentException ( ( ( ( "Attempting<sp>to<sp>split<sp>on<sp>EndRow<sp>" + ( extent . getEndRow ( ) ) ) + "<sp>for<sp>" ) + ( extent ) ) ) ; } if ( ( sp != null ) && ( ( sp . length ) > ( tableConfiguration . getAsBytes ( Property . TABLE_MAX_END_ROW_SIZE ) ) ) ) { java . lang . String msg = ( ( "Cannot<sp>split<sp>tablet<sp>" + ( extent ) ) + ",<sp>selected<sp>split<sp>point<sp>too<sp>long.<sp>Length<sp>:<sp>" ) + ( sp . length ) ; org . apache . accumulo . tserver . tablet . Tablet . log . warn ( msg ) ; throw new java . io . IOException ( msg ) ; } if ( extent . isRootTablet ( ) ) { java . lang . String msg = ",<sp>selected<sp>split<sp>point<sp>too<sp>long.<sp>Length<sp>:<sp>" 0 ; org . apache . accumulo . tserver . tablet . Tablet . log . warn ( msg ) ; throw new java . lang . RuntimeException ( msg ) ; } try { initiateClose ( true , false , false ) ; } catch ( java . lang . IllegalStateException ise ) { org . apache . accumulo . tserver . tablet . Tablet . log . debug ( "File<sp>{}<sp>not<sp>splitting<sp>:<sp>{}" , extent , ise . getMessage ( ) ) ; return null ; } java . util . Map < org . apache . accumulo . server . fs . FileRef , org . apache . accumulo . server . util . FileUtil . FileInfo > firstAndLastRows = org . apache . accumulo . server . util . FileUtil . tryToGetFirstAndLastRows ( context , getDatafileManager ( ) . getFiles ( ) ) ; synchronized ( this ) { java . util . TreeMap < org . apache . accumulo . core . dataImpl . KeyExtent , org . apache . accumulo . tserver . tablet . TabletData > newTablets = new java . util . TreeMap ( ) ; long t1 = java . lang . System . currentTimeMillis ( ) ; org . apache . accumulo . tserver . tablet . SplitRowSpec splitPoint ; if ( sp == null ) splitPoint = findSplitRow ( getDatafileManager ( ) . getFiles ( ) ) ; else { org . apache . hadoop . io . Text tsp = new org . apache . hadoop . io . Text ( sp ) ; splitPoint = new org . apache . accumulo . tserver . tablet . SplitRowSpec ( org . apache . accumulo . server . util . FileUtil . estimatePercentageLTE ( context , tabletDirectory , extent . getPrevEndRow ( ) , extent . getEndRow ( ) , org . apache . accumulo . server . util . FileUtil . toPathStrings ( getDatafileManager ( ) . getFiles ( ) ) , tsp ) , tsp ) ; } if ( ( splitPoint == null ) || ( ( splitPoint . row ) == null ) ) { org . apache . accumulo . tserver . tablet . Tablet . log . info ( "had<sp>to<sp>abort<sp>split<sp>because<sp>splitRow<sp>was<sp>null" ) ; closeState = org . apache . accumulo . tserver . tablet . Tablet . CloseState . OPEN ; return null ; } closeState = org . apache . accumulo . tserver . tablet . Tablet . CloseState . CLOSING ; completeClose ( true , false ) ; org . apache . hadoop . io . Text midRow = splitPoint . row ; double splitRatio = splitPoint . splitRatio ; org . apache . accumulo . core . dataImpl . KeyExtent low = new org . apache . accumulo . core . dataImpl . KeyExtent ( extent . getTableId ( ) , midRow , extent . getPrevEndRow ( ) ) ; org . apache . accumulo . core . dataImpl . KeyExtent high = new org . apache . accumulo . core . dataImpl . KeyExtent ( extent . getTableId ( ) , extent . getEndRow ( ) , midRow ) ; java . lang . String lowDirectory = org . apache . accumulo . tserver . tablet . Tablet . createTabletDirectory ( context , getTabletServer ( ) . getFileSystem ( ) , extent . getTableId ( ) , midRow ) ; java . util . SortedMap < org . apache . accumulo . server . fs . FileRef , org . apache . accumulo . core . metadata . schema . DataFileValue
org . junit . Assert . assertEquals ( 12 , parts . length )
testOtherObjectsSameProperties ( ) { com . ning . atlas . spi . Identity id = com . ning . atlas . spi . Identity . root ( ) . createChild ( "test" , "0" ) . createChild ( "waffle" , "9" ) ; com . ning . atlas . space . BaseSpaceTest . Thing t = new com . ning . atlas . space . BaseSpaceTest . Thing ( ) ; t . setName ( "Freddy" ) ; t . setAgeOfPetDog ( 14 ) ; space . store ( id , t ) ; com . ning . atlas . space . BaseSpaceTest . NameOnly no = space . get ( id , com . ning . atlas . space . BaseSpaceTest . NameOnly . class , Missing . RequireAll ) . getValue ( ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; }
org . junit . Assert . assertThat ( no . getName ( ) , org . hamcrest . CoreMatchers . equalTo ( t . getName ( ) ) )
toVOCatalogEntry_NullInput ( ) { org . oscm . internal . vo . VOCatalogEntry entry = org . oscm . serviceprovisioningservice . assembler . CatalogEntryAssembler . toVOCatalogEntry ( null , facade ) ; "<AssertPlaceHolder>" ; } toVOCatalogEntry ( org . oscm . domobjects . CatalogEntry , org . oscm . i18nservice . bean . LocalizerFacade ) { return org . oscm . serviceprovisioningservice . assembler . CatalogEntryAssembler . toVOCatalogEntry ( catalogEntry , facade , PerformanceHint . ALL_FIELDS ) ; }
org . junit . Assert . assertNull ( entry )
headerFoundButNoValidContentTest ( ) { when ( headerMock . getValue ( ) ) . thenReturn ( "AX4" ) ; when ( httpResponseMock . getFirstHeader ( "Retry-After" ) ) . thenReturn ( headerMock ) ; de . bytefish . fcmjava . client . utils . OutParameter < java . time . Duration > result = new de . bytefish . fcmjava . client . utils . OutParameter ( ) ; boolean success = de . bytefish . fcmjava . client . http . apache . utils . RetryHeaderUtils . tryDetermineRetryDelay ( httpResponseMock , result ) ; "<AssertPlaceHolder>" ; } tryDetermineRetryDelay ( org . apache . http . HttpResponse , de . bytefish . fcmjava . client . utils . OutParameter ) { try { return de . bytefish . fcmjava . client . http . apache . utils . RetryHeaderUtils . internalTryDetermineRetryDelay ( httpResponse , result ) ; } catch ( java . lang . Exception e ) { return false ; } }
org . junit . Assert . assertEquals ( false , success )
testSaveEntityWithId ( ) { com . github . jmkgreen . morphia . testdaos . HotelDAO hotelDAO = new com . github . jmkgreen . morphia . testdaos . HotelDAO ( morphia , mongo ) ; com . github . jmkgreen . morphia . testmodel . Hotel borg = com . github . jmkgreen . morphia . testmodel . Hotel . create ( ) ; borg . setName ( "Hotel<sp>Borg" ) ; borg . setStars ( 4 ) ; hotelDAO . save ( borg ) ; com . github . jmkgreen . morphia . testmodel . Hotel hotelLoaded = hotelDAO . get ( borg . getId ( ) ) ; hotelLoaded . setStars ( 5 ) ; hotelDAO . save ( hotelLoaded ) ; com . github . jmkgreen . morphia . testmodel . Hotel hotelReloaded = hotelDAO . get ( borg . getId ( ) ) ; "<AssertPlaceHolder>" ; } getStars ( ) { return stars ; }
org . junit . Assert . assertEquals ( 5 , hotelReloaded . getStars ( ) )
test5a ( ) { cc . redberry . rings . IntegersZp64 domain = new cc . redberry . rings . IntegersZp64 ( cc . redberry . rings . primes . SmallPrimes . nextPrime ( 66 ) ) ; java . lang . String [ ] vars = new java . lang . String [ ] { "a" , "b" , "c" } ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 [ ] factors = new cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 [ ] { cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . parse ( "a^15*b*c^2<sp>+<sp>2*a*b^4<sp>-<sp>3*b<sp>+<sp>2<sp>+<sp>b^2*a<sp>-<sp>b^4" , domain , vars ) , cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 . parse ( "a^5*b^6*c^3<sp>+<sp>a*b^2<sp>-<sp>3*b^2<sp>+<sp>b<sp>+<sp>2<sp>-<sp>a^3*b^6" , domain , vars ) } ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 base = cc . redberry . rings . poly . multivar . HenselLiftingTest . multiply ( factors ) ; assert cc . redberry . rings . poly . multivar . MultivariateGCD . PolynomialGCD ( factors ) . isConstant ( ) ; cc . redberry . rings . poly . multivar . HenselLifting . lEvaluation evaluation = new cc . redberry . rings . poly . multivar . HenselLifting . lEvaluation ( base . nVariables , new long [ ] { 1 , 2 } , domain , base . ordering ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 [ ] biFactors = evaluation . evaluateFrom ( factors , 2 ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 lc = base . lc ( 0 ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 lcCorrection = evaluation . evaluateFrom ( lc , 2 ) ; System . out . println ( java . util . Arrays . stream ( biFactors ) . map ( ( p ) -> p . lc ( 0 ) ) . reduce ( base . createOne ( ) , ( a , b ) -> a . clone ( ) . multiply ( b ) ) ) ; System . out . println ( evaluation . evaluateFrom ( base . lc ( 0 ) , 2 ) ) ; for ( cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 factor : biFactors ) { cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 r = cc . redberry . rings . poly . multivar . MultivariateDivision . divideExact ( lcCorrection , factor . lc ( 0 ) ) ; factor . multiply ( r ) ; } cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 tmp = base . clone ( ) . multiply ( cc . redberry . rings . poly . PolynomialMethods . polyPow ( lc , ( ( biFactors . length ) - 1 ) , true ) ) ; cc . redberry . rings . poly . multivar . HenselLifting . multivariateLift0 ( tmp , biFactors , cc . redberry . rings . util . ArraysUtil . arrayOf ( lc , biFactors . length ) , evaluation , base . degrees ( ) , 2 ) ; for ( cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 factor : biFactors ) factor . set ( cc . redberry . rings . poly . multivar . HenselLifting . primitivePart ( factor ) ) ; System . out . println ( java . util . Arrays . toString ( factors ) ) ; System . out . println ( java . util . Arrays . toString ( biFactors ) ) ; "<AssertPlaceHolder>" ; } multiply ( Poly extends cc . redberry . rings . poly . multivar . AMultivariatePolynomial ) { factory . assertSameCoefficientRingWith ( oth ) ; if ( isTrivial ( ) ) return cc . redberry . rings . poly . multivar . Ideal . create ( cc . redberry . rings . poly . multivar . Collections . singletonList ( oth ) , ordering ) ; if ( oth . isZero ( ) ) return cc . redberry . rings . poly . multivar . Ideal . trivial ( oth , ordering ) ; if ( ( oth . isOne ( ) ) || ( this . isEmpty ( ) ) ) return this ; return new cc . redberry . rings . poly . multivar . Ideal ( canonicalize ( groebnerBasis . stream ( ) . map ( ( p ) -> p . clone ( ) . multiply ( oth ) ) . collect ( java . util . stream . Collectors . toList ( ) ) ) ) ; }
org . junit . Assert . assertEquals ( base , cc . redberry . rings . poly . multivar . HenselLiftingTest . multiply ( biFactors ) )
testMatchingA ( ) { java . io . File f1 = new java . io . File ( "dir" ) ; java . io . File f2 = new java . io . File ( "dir/eSTARt" ) ; java . io . File f3 = new java . io . File ( "dir/notme" ) ; java . io . File f4 = new java . io . File ( "./" ) ; com . epimorphics . lda . support . Glob g = new com . epimorphics . lda . support . Glob ( new com . epimorphics . lda . support . tests . TestGlob . FakeFSI ( f1 , f2 , f3 , f4 ) ) ; com . epimorphics . lda . support . tests . Set < java . io . File > files = new com . epimorphics . lda . support . tests . HashSet < java . io . File > ( g . filesMatching ( "dir/e*t" ) ) ; "<AssertPlaceHolder>" ; } set ( T [ ] ) { return new com . epimorphics . util . HashSet < T > ( com . epimorphics . util . CollectionUtils . list ( elements ) ) ; }
org . junit . Assert . assertEquals ( com . epimorphics . util . CollectionUtils . set ( f2 ) , files )
testEndTime ( ) { final javax . ejb . ScheduleExpression expr = new javax . ejb . ScheduleExpression ( ) . dayOfMonth ( 20 ) . dayOfWeek ( "sat" ) . start ( new java . util . Date ( 0 ) ) ; final org . apache . openejb . core . timer . EJBCronTrigger trigger = new org . apache . openejb . core . timer . EJBCronTrigger ( expr ) ; java . util . Calendar calendar = new java . util . GregorianCalendar ( 2008 , 0 , 4 ) ; trigger . setEndTime ( calendar . getTime ( ) ) ; calendar = new java . util . GregorianCalendar ( 2008 , 0 , 1 ) ; "<AssertPlaceHolder>" ; } getFireTimeAfter ( java . util . Date ) { org . apache . openejb . core . timer . EJBCronTrigger . log . debug ( ( "start<sp>to<sp>getFireTimeAfter:" + afterTime ) ) ; final java . util . Calendar calendar = new java . util . GregorianCalendar ( timezone ) ; calendar . setFirstDayOfWeek ( Calendar . SUNDAY ) ; if ( ( ( getStartTime ( ) ) != null ) && ( getStartTime ( ) . after ( afterTime ) ) ) { calendar . setTime ( getStartTime ( ) ) ; } else { calendar . setTime ( afterTime ) ; calendar . add ( Calendar . SECOND , 1 ) ; } final java . util . Calendar stopCalendar = new java . util . GregorianCalendar ( timezone ) ; if ( ( getEndTime ( ) ) != null ) { stopCalendar . setTime ( getEndTime ( ) ) ; } else { final int stopYear = ( calendar . get ( Calendar . YEAR ) ) + 100 ; stopCalendar . set ( Calendar . YEAR , stopYear ) ; } int currentFieldIndex = 0 ; while ( ( currentFieldIndex <= 6 ) && ( calendar . before ( stopCalendar ) ) ) { final org . apache . openejb . core . timer . EJBCronTrigger . FieldExpression expr = expressions [ currentFieldIndex ] ; java . lang . Integer value = expr . getNextValue ( calendar ) ; else if ( value != null ) { final int oldValue = calendar . get ( expr . field ) ; if ( oldValue != value ) { if ( ( currentFieldIndex == 3 ) && ( ! ( ( expressions [ 2 ] ) instanceof org . apache . openejb . core . timer . EJBCronTrigger . AsteriskExpression ) ) ) { currentFieldIndex ++ ; } else { calendar . set ( expr . field , value ) ; resetFields ( calendar , expr . field , false ) ; currentFieldIndex ++ ; } } else { currentFieldIndex ++ ; } } else { org . apache . openejb . core . timer . EJBCronTrigger . log . debug ( ( "end<sp>of<sp>getFireTimeAfter,<sp>result<sp>is:" + null ) ) ; return null ; } } org . apache . openejb . core . timer . EJBCronTrigger . log . debug ( ( "end<sp>of<sp>getFireTimeAfter,<sp>result<sp>is:" + ( calendar . before ( stopCalendar ) ? calendar . getTime ( ) : null ) ) ) ; return calendar . before ( stopCalendar ) ? calendar . getTime ( ) : null ; }
org . junit . Assert . assertNull ( trigger . getFireTimeAfter ( calendar . getTime ( ) ) )
fileSizeDecreasesAfterResize ( ) { java . io . File sourceFile = new java . io . File ( "test-resources/Thumbnailator/grid.png" ) ; java . io . File f = new java . io . File ( "test-resources/Thumbnailator/tmp-grid.png" ) ; net . coobird . thumbnailator . TestUtils . copyFile ( sourceFile , f ) ; long fileSizeBefore = f . length ( ) ; net . coobird . thumbnailator . Thumbnails . of ( f ) . size ( 100 , 100 ) . toFile ( f ) ; long fileSizeAfter = f . length ( ) ; f . delete ( ) ; "<AssertPlaceHolder>" ; } size ( int , int ) { if ( width < 0 ) { throw new java . lang . IllegalArgumentException ( "Width<sp>must<sp>be<sp>greater<sp>than<sp>0." ) ; } if ( height < 0 ) { throw new java . lang . IllegalArgumentException ( "Height<sp>must<sp>be<sp>greater<sp>than<sp>0." ) ; } this . width = width ; this . height = height ; return this ; }
org . junit . Assert . assertTrue ( ( fileSizeAfter < fileSizeBefore ) )
testFrameRowsUnboundedPrecedingCurrentRow ( ) { java . lang . String sqlText = java . lang . String . format ( ( "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 6 + ( "min(salary)<sp>over<sp>(Partition<sp>by<sp>dept<sp>ORDER<sp>BY<sp>salary<sp>rows<sp>between<sp>unbounded<sp>preceding<sp>and<sp>current<sp>row)<sp>as<sp>minsal<sp>" + "from<sp>%s<sp>--SPLICE-PROPERTIES<sp>useSpark<sp>=<sp>%s<sp>\n<sp>order<sp>by<sp>empnum" ) ) , this . getTableReference ( com . splicemachine . derby . impl . sql . execute . operations . WindowFunctionIT . EMPTAB ) , useSpark ) ; java . sql . ResultSet rs = com . splicemachine . derby . impl . sql . execute . operations . WindowFunctionIT . methodWatcher . executeQuery ( sqlText ) ; java . lang . String expected = "EMPNUM<sp>|DEPT<sp>|SALARY<sp>|MINSAL<sp>|\n" + ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( "------------------------------\n" + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 1 ) + "<sp>20<sp>|<sp>1<sp>|<sp>75000<sp>|<sp>50000<sp>|\n" ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 2 ) + "<sp>40<sp>|<sp>2<sp>|<sp>52000<sp>|<sp>51000<sp>|\n" ) + "<sp>44<sp>|<sp>2<sp>|<sp>52000<sp>|<sp>51000<sp>|\n" ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 7 ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 9 ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 8 ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 4 ) + "<sp>80<sp>|<sp>3<sp>|<sp>79000<sp>|<sp>55000<sp>|\n" ) + "<sp>90<sp>|<sp>2<sp>|<sp>51000<sp>|<sp>51000<sp>|\n" ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 5 ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 3 ) + "<sp>50<sp>|<sp>1<sp>|<sp>52000<sp>|<sp>50000<sp>|\n" 0 ) ; "<AssertPlaceHolder>" ; rs . close ( ) ; } toStringUnsorted ( com . splicemachine . homeless . ResultSet ) { return com . splicemachine . homeless . TestUtils . FormattedResult . ResultFactory . convert ( "" , rs , false ) . toString ( ) . trim ( ) ; }
org . junit . Assert . assertEquals ( ( ( "from<sp>%s<sp>--SPLICE-PROPERTIES<sp>useSpark<sp>=<sp>%s<sp>\n<sp>order<sp>by<sp>empnum" 0 + sqlText ) + "from<sp>%s<sp>--SPLICE-PROPERTIES<sp>useSpark<sp>=<sp>%s<sp>\n<sp>order<sp>by<sp>empnum" 0 ) , expected , TestUtils . FormattedResult . ResultFactory . toStringUnsorted ( rs ) )
createBranchFromCommitId_theHeadOfTheNewBranchShouldEqualToTheSpecifiedCommit ( ) { writeSomethingToCache ( ) ; org . eclipse . jgit . lib . AnyObjectId commit = commitToMaster ( ) ; com . beijunyi . parallelgit . utils . BranchUtils . createBranch ( "test_branch" , commit , repo ) ; "<AssertPlaceHolder>" ; } getHeadCommit ( java . lang . String , com . beijunyi . parallelgit . utils . Repository ) { com . beijunyi . parallelgit . utils . Ref ref = com . beijunyi . parallelgit . utils . RefUtils . getBranchRef ( name , repo ) ; return com . beijunyi . parallelgit . utils . CommitUtils . getCommit ( ref . getObjectId ( ) , repo ) ; }
org . junit . Assert . assertEquals ( commit , com . beijunyi . parallelgit . utils . BranchUtils . getHeadCommit ( "test_branch" , repo ) )
clientRequestOneDataFrame ( ) { boostrapEnv ( 1 , 1 , 1 ) ; final java . lang . String text = "hello<sp>world" ; final io . netty . buffer . ByteBuf content = io . netty . buffer . Unpooled . copiedBuffer ( text . getBytes ( ) ) ; final io . netty . handler . codec . http . FullHttpRequest request = new io . netty . handler . codec . http . DefaultFullHttpRequest ( io . netty . handler . codec . http . HttpVersion . HTTP_1_1 , io . netty . handler . codec . http . HttpMethod . GET , "/some/path/resource2" , content , true ) ; try { io . netty . handler . codec . http . HttpHeaders httpHeaders = request . headers ( ) ; httpHeaders . setInt ( HttpConversionUtil . ExtensionHeaderNames . STREAM_ID . text ( ) , 3 ) ; httpHeaders . setInt ( HttpHeaderNames . CONTENT_LENGTH , text . length ( ) ) ; httpHeaders . setShort ( HttpConversionUtil . ExtensionHeaderNames . STREAM_WEIGHT . text ( ) , ( ( short ) ( 16 ) ) ) ; final io . netty . handler . codec . http2 . Http2Headers http2Headers = new io . netty . handler . codec . http2 . DefaultHttp2Headers ( ) . method ( new io . netty . util . AsciiString ( "GET" ) ) . path ( new io . netty . util . AsciiString ( "/some/path/resource2" ) ) ; io . netty . handler . codec . http2 . Http2TestUtil . runInChannel ( clientChannel , new io . netty . handler . codec . http2 . Http2TestUtil . Http2Runnable ( ) { @ io . netty . handler . codec . http2 . Override public void run ( ) throws io . netty . handler . codec . http2 . Http2Exception { clientHandler . encoder ( ) . writeHeaders ( ctxClient ( ) , 3 , http2Headers , 0 , false , newPromiseClient ( ) ) ; clientHandler . encoder ( ) . writeData ( ctxClient ( ) , 3 , content . retainedDuplicate ( ) , 0 , true , newPromiseClient ( ) ) ; clientChannel . flush ( ) ; } } ) ; awaitRequests ( ) ; org . mockito . ArgumentCaptor < io . netty . handler . codec . http . FullHttpMessage > requestCaptor = org . mockito . ArgumentCaptor . forClass ( io . netty . handler . codec . http . FullHttpMessage . class ) ; verify ( serverListener ) . messageReceived ( requestCaptor . capture ( ) ) ; capturedRequests = requestCaptor . getAllValues ( ) ; "<AssertPlaceHolder>" ; } finally { request . release ( ) ; } } get ( java . lang . CharSequence ) { return get0 ( name ) ; }
org . junit . Assert . assertEquals ( request , capturedRequests . get ( 0 ) )
testClone ( ) { org . apache . phoenix . expression . Expression inputArg = org . apache . phoenix . expression . LiteralExpression . newConstant ( PIntegerArray . INSTANCE . getSqlType ( ) , PInteger . INSTANCE ) ; java . util . List < org . apache . phoenix . expression . Expression > args = com . google . common . collect . Lists . newArrayList ( inputArg ) ; org . apache . phoenix . expression . function . ExternalSqlTypeIdFunction externalIdFunction = new org . apache . phoenix . expression . function . ExternalSqlTypeIdFunction ( args ) ; org . apache . phoenix . expression . function . ScalarFunction clone = externalIdFunction . clone ( args ) ; "<AssertPlaceHolder>" ; } clone ( java . util . List ) { return new org . apache . phoenix . expression . IsNullExpression ( children , this . isNegate ( ) ) ; }
org . junit . Assert . assertEquals ( externalIdFunction , clone )
testConstructGraphSerializerBlankNode ( ) { java . lang . String query = "select<sp>?x<sp>where<sp>{<sp>_:b<sp><uri:talksTo><sp>?x.<sp>_:b<sp><uri:worksAt><sp>?y<sp>}" ; org . eclipse . rdf4j . query . parser . sparql . SPARQLParser parser = new org . eclipse . rdf4j . query . parser . sparql . SPARQLParser ( ) ; org . eclipse . rdf4j . query . parser . ParsedQuery pq = parser . parseQuery ( query , null ) ; java . util . List < org . eclipse . rdf4j . query . algebra . StatementPattern > patterns = org . eclipse . rdf4j . query . algebra . helpers . StatementPatternCollector . process ( pq . getTupleExpr ( ) ) ; org . apache . rya . indexing . pcj . fluo . app . ConstructGraph graph = new org . apache . rya . indexing . pcj . fluo . app . ConstructGraph ( patterns ) ; java . lang . String constructString = org . apache . rya . indexing . pcj . fluo . app . ConstructGraphSerializer . toConstructString ( graph ) ; org . apache . rya . indexing . pcj . fluo . app . ConstructGraph deserialized = org . apache . rya . indexing . pcj . fluo . app . ConstructGraphSerializer . toConstructGraph ( constructString ) ; "<AssertPlaceHolder>" ; } toConstructGraph ( java . lang . String ) { java . util . Set < org . apache . rya . indexing . pcj . fluo . app . ConstructProjection > projections = new java . util . HashSet ( ) ; java . lang . String [ ] spStrings = graphString . split ( org . apache . rya . indexing . pcj . fluo . app . ConstructGraphSerializer . SP_DELIM ) ; for ( java . lang . String sp : spStrings ) { projections . add ( new org . apache . rya . indexing . pcj . fluo . app . ConstructProjection ( org . apache . rya . indexing . pcj . fluo . app . FluoStringConverter . toStatementPattern ( sp ) ) ) ; } return new org . apache . rya . indexing . pcj . fluo . app . ConstructGraph ( projections ) ; }
org . junit . Assert . assertEquals ( graph , deserialized )
testEquals_wrongClass ( ) { "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == o ) { return true ; } if ( ( o == null ) || ( ( getClass ( ) ) != ( o . getClass ( ) ) ) ) { return false ; } org . esa . beam . binning . operator . BinningOp . BandConfiguration that = ( ( org . esa . beam . binning . operator . BinningOp . BandConfiguration ) ( o ) ) ; if ( ( index ) != null ? ! ( index . equals ( that . index ) ) : ( that . index ) != null ) { return false ; } if ( ( maxValue ) != null ? ! ( maxValue . equals ( that . maxValue ) ) : ( that . maxValue ) != null ) { return false ; } if ( ( minValue ) != null ? ! ( minValue . equals ( that . minValue ) ) : ( that . minValue ) != null ) { return false ; } if ( ( name ) != null ? ! ( name . equals ( that . name ) ) : ( that . name ) != null ) { return false ; } return true ; }
org . junit . Assert . assertFalse ( config . equals ( new java . lang . Double ( 4 ) ) )
testSessionSerialization ( ) { it . redhat . hacep . rules . TestSerializedRetractRules . logger . info ( "Start<sp>test<sp>serialized<sp>rules" ) ; java . lang . System . setProperty ( "grid.buffer" , "10" ) ; it . redhat . hacep . cluster . RulesConfigurationTestImpl rulesConfigurationTest = RulesConfigurationTestImpl . RulesTestBuilder . buildRulesWithGamePlayRetract ( ) ; it . redhat . hacep . configuration . RulesManager rulesManager = new it . redhat . hacep . configuration . RulesManager ( rulesConfigurationTest ) ; rulesManager . start ( null , null , null ) ; reset ( outcomesChannel ) ; org . kie . api . runtime . KieSession kieSession = rulesManager . newKieSession ( ) ; kieSession . registerChannel ( "outcomes" , outcomesChannel ) ; kieSession . insert ( generateFactTenSecondsAfter ( 1 ) ) ; kieSession . fireAllRules ( ) ; verify ( outcomesChannel , times ( 1 ) ) . send ( any ( ) ) ; verifyNoMoreInteractions ( outcomesChannel ) ; reset ( outcomesChannel ) ; byte [ ] kieSessionBytes = rulesManager . serialize ( kieSession ) ; "<AssertPlaceHolder>" ; kieSession . dispose ( ) ; org . kie . api . runtime . KieSession kieSessionDeserialized = rulesManager . deserializeOrCreate ( kieSessionBytes ) ; kieSessionDeserialized . registerChannel ( "outcomes" , outcomesChannel ) ; kieSessionDeserialized . insert ( generateFactTenSecondsAfter ( 1 ) ) ; kieSessionDeserialized . fireAllRules ( ) ; verify ( outcomesChannel , times ( 1 ) ) . send ( any ( ) ) ; verifyNoMoreInteractions ( outcomesChannel ) ; it . redhat . hacep . rules . TestSerializedRetractRules . logger . info ( "End<sp>test<sp>serialized<sp>rules" ) ; rulesManager . stop ( ) ; } serialize ( org . kie . api . runtime . KieSession ) { return this . serialize ( this . kieContainer , kieSession ) ; }
org . junit . Assert . assertTrue ( ( ( kieSessionBytes . length ) > 0 ) )