input stringlengths 28 18.7k | output stringlengths 39 1.69k |
|---|---|
shouldFindAll ( ) { final com . couchbase . client . java . bucket . BucketManager bucketManager = bucket . bucketManager ( ) ; final com . couchbase . client . java . view . View view = com . couchbase . client . java . view . DefaultView . create ( "all" , org . apache . commons . io . IOUtils . toString ( getClass ( ) . getResourceAsStream ( "/all.js" ) ) ) ; final com . couchbase . client . java . view . DesignDocument document = com . couchbase . client . java . view . DesignDocument . create ( "person" , com . google . common . collect . ImmutableList . of ( view ) ) ; try { bucketManager . upsertDesignDocument ( document ) ; repository . save ( person ) ; final java . lang . Iterable < com . github . jloisel . reactive . repository . couchbase . it . Person > found = repository . findAll ( ) ; "<AssertPlaceHolder>" ; } finally { bucketManager . removeDesignDocument ( document . name ( ) ) ; } } findAll ( ) { return blocking ( async . findAll ( ) . toList ( ) ) . single ( ) ; } | org . junit . Assert . assertEquals ( com . google . common . collect . ImmutableList . of ( person ) , found ) |
testCharAt ( ) { java . lang . String str = "http://owlapi.sourceforge.net#ABC" ; org . semanticweb . owlapi . api . test . ontology . IRI iri = IRI ( "http://owlapi.sourceforge.net#" , "ABC" ) ; for ( int i = 0 ; i < ( str . length ( ) ) ; i ++ ) { "<AssertPlaceHolder>" ; } } charAt ( int ) { if ( index < 0 ) { throw new java . lang . IndexOutOfBoundsException ( java . lang . Integer . toString ( index ) ) ; } if ( index < ( namespace . length ( ) ) ) { return namespace . charAt ( index ) ; } return remainder . charAt ( ( index - ( namespace . length ( ) ) ) ) ; } | org . junit . Assert . assertEquals ( str . charAt ( i ) , iri . charAt ( i ) ) |
hasNextSucceeds ( tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Query ) { final java . lang . String connectionString = "testString" ; com . microsoft . azure . sdk . iot . service . jobs . JobClient testJobClient = com . microsoft . azure . sdk . iot . service . jobs . JobClient . createFromConnectionString ( connectionString ) ; new tests . unit . com . microsoft . azure . sdk . iot . service . jobs . NonStrictExpectations ( ) { { tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Deencapsulation . newInstance ( tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Query . class , new java . lang . Class [ ] { java . lang . String . class , tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Integer . class , tests . unit . com . microsoft . azure . sdk . iot . service . jobs . QueryType . class } , anyString , anyInt , QueryType . DEVICE_JOB ) ; result = mockedQuery ; tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Deencapsulation . invoke ( mockedQuery , "hasNext" ) ; result = true ; } } ; tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Query testQuery = testJobClient . queryDeviceJob ( tests . unit . com . microsoft . azure . sdk . iot . service . jobs . JobClientTest . VALID_SQL_QUERY ) ; boolean result = testJobClient . hasNextJob ( testQuery ) ; new tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Verifications ( ) { { tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Deencapsulation . invoke ( mockedQuery , "sendQueryRequest" , new java . lang . Class [ ] { com . microsoft . azure . sdk . iot . service . IotHubConnectionString . class , java . net . URL . class , com . microsoft . azure . sdk . iot . service . transport . http . HttpMethod . class , tests . unit . com . microsoft . azure . sdk . iot . service . jobs . Long . class } , any , any , HttpMethod . POST , any ) ; times = 1 ; } } ; "<AssertPlaceHolder>" ; } hasNextJob ( com . microsoft . azure . sdk . iot . service . jobs . Query ) { if ( query == null ) { throw new java . lang . IllegalArgumentException ( "Query<sp>cannot<sp>be<sp>null" ) ; } return query . hasNext ( ) ; } | org . junit . Assert . assertTrue ( result ) |
CollisionWithMergeInlineAbove2RemoveOneCollisonNode ( ) { org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue hash98304_obj1 = new org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue ( 1 , 98304 ) ; org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue hash98304_obj2 = new org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue ( 2 , 98304 ) ; org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue hash268435456_obj3 = new org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue ( 3 , 268435456 ) ; org . eclipse . imp . pdb . facts . util . ImmutableSet < org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue > xs = org . eclipse . imp . pdb . facts . util . TrieSet_5Bits . of ( hash98304_obj1 , hash268435456_obj3 , hash98304_obj2 ) . __remove ( hash98304_obj2 ) ; org . eclipse . imp . pdb . facts . util . ImmutableSet < org . eclipse . imp . pdb . test . persistent . TrieSetTests . DummyValue > ys = org . eclipse . imp . pdb . facts . util . TrieSet_5Bits . of ( hash98304_obj1 , hash268435456_obj3 ) ; "<AssertPlaceHolder>" ; } __remove ( K ) { final int keyHash = key . hashCode ( ) ; final org . eclipse . imp . pdb . facts . util . TrieSet_5Bits . SetResult < K > details = org . eclipse . imp . pdb . facts . util . TrieSet_5Bits . SetResult . unchanged ( ) ; final org . eclipse . imp . pdb . facts . util . TrieSet_5Bits . CompactSetNode < K > newRootNode = rootNode . removed ( null , key , org . eclipse . imp . pdb . facts . util . TrieSet_5Bits . transformHashCode ( keyHash ) , 0 , details ) ; if ( details . isModified ( ) ) { return new org . eclipse . imp . pdb . facts . util . TrieSet_5Bits < K > ( newRootNode , ( ( hashCode ) - keyHash ) , ( ( cachedSize ) - 1 ) ) ; } return this ; } | org . junit . Assert . assertEquals ( xs , ys ) |
testGetTempFactor ( ) { org . openscience . cdk . interfaces . IPDBAtom atom = ( ( org . openscience . cdk . interfaces . IPDBAtom ) ( newChemObject ( ) ) ) ; atom . setSymbol ( "C" ) ; atom . setTempFactor ( 0.0 ) ; "<AssertPlaceHolder>" ; } getTempFactor ( ) { return tempFactor ; } | org . junit . Assert . assertEquals ( atom . getTempFactor ( ) , 0.0 , 1 ) |
testGetRequestHeaders ( ) { System . out . println ( "getRequestHeaders" ) ; com . atlantbh . jmeter . plugins . oauth . OAuthSampler instance = new com . atlantbh . jmeter . plugins . oauth . OAuthSampler ( ) ; java . lang . String expResult = "" ; java . lang . String result = instance . getRequestHeaders ( ) ; "<AssertPlaceHolder>" ; } getRequestHeaders ( ) { return getPropertyAsString ( com . atlantbh . jmeter . plugins . oauth . OAuthSampler . REQUEST_HEADERS ) ; } | org . junit . Assert . assertEquals ( expResult , result ) |
testSetVolumeSize ( ) { com . iwave . ext . netapp . Volume vol = new com . iwave . ext . netapp . Volume ( com . iwave . ext . netapp . VolumeTest . server , com . iwave . ext . netapp . VolumeTest . VOL_NAME ) ; java . lang . String size = vol . setVolumeSize ( com . iwave . ext . netapp . VolumeTest . NEW_VOL_SIZE ) ; "<AssertPlaceHolder>" ; } setVolumeSize ( java . lang . Long ) { this . volumeSize = _volumeSize ; } | org . junit . Assert . assertEquals ( com . iwave . ext . netapp . VolumeTest . NEW_VOL_SIZE , size ) |
shouldReturnHeliosHostnameWithNoDomainForSyslogAppender ( ) { environmentVariables . set ( LoggingConfigurator . SPOTIFY_HOSTNAME , "hostname" ) ; com . spotify . logging . LoggingConfigurator . configureSyslogDefaults ( "idnet" ) ; "<AssertPlaceHolder>" ; } getLoggingContextHostnameProperty ( ) { final ch . qos . logback . classic . Logger accessPointLogger = ( ( ch . qos . logback . classic . Logger ) ( org . slf4j . LoggerFactory . getLogger ( "logger" ) ) ) ; final ch . qos . logback . classic . LoggerContext loggerContext = accessPointLogger . getLoggerContext ( ) ; return loggerContext . getProperty ( "hostname" ) ; } | org . junit . Assert . assertEquals ( "hostname" , getLoggingContextHostnameProperty ( ) ) |
testH11 ( ) { java . lang . Object [ ] input = new java . lang . Object [ ] { 1.1 , true , "ABC" } ; java . lang . String result = org . terasoluna . gfw . web . el . Functions . h ( input ) ; "<AssertPlaceHolder>" ; } h ( java . lang . Object ) { return org . terasoluna . gfw . web . util . HtmlEscapeUtils . htmlEscape ( input ) ; } | org . junit . Assert . assertThat ( result , org . hamcrest . CoreMatchers . is ( java . util . Arrays . toString ( ( ( java . lang . Object [ ] ) ( input ) ) ) ) ) |
a_loadWMSGetFeatureReaderASTextPlainTest ( ) { org . geosdi . geoplatform . connector . bridge . implementor . GPWMSGetFeatureInfoReader < ? > textPlainGetFeatureReader = org . geosdi . geoplatform . connector . wms . bridge . store . GPWMSGetFeatureInfoReaderStoreTest . store . getImplementorByKey ( org . geosdi . geoplatform . connector . wms . bridge . store . TEXT_PLAIN ) ; "<AssertPlaceHolder>" ; org . geosdi . geoplatform . connector . wms . bridge . store . GPWMSGetFeatureInfoReaderStoreTest . logger . info ( "################################TEXT_PLAIN_READER<sp>:<sp>{}\n" , textPlainGetFeatureReader ) ; } getImplementorByKey ( org . geosdi . geoplatform . support . bridge . implementor . GPImplementor . GPImplementorKey ) { checkArgument ( ( key != null ) , "The<sp>Parameter<sp>key<sp>must<sp>not<sp>be<sp>null" ) ; org . geosdi . geoplatform . connector . bridge . store . GPWMSGetFeatureInfoReaderStore . logger . trace ( "@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@Try<sp>to<sp>find<sp>WMSGetFeatureInfoReader<sp>with<sp>Key<sp>:<sp>{}\n" , key ) ; return org . geosdi . geoplatform . connector . bridge . store . GPWMSGetFeatureInfoReaderStore . wmsGetFeatureInfoReaders . get ( key ) ; } | org . junit . Assert . assertNotNull ( textPlainGetFeatureReader ) |
testDateIsNull ( ) { when ( jsonParser . getCurrentToken ( ) ) . thenReturn ( com . fasterxml . jackson . core . JsonToken . VALUE_NULL ) ; deserializedDate = deserializer . deserialize ( jsonParser , deserializationContext ) ; "<AssertPlaceHolder>" ; } deserialize ( com . fasterxml . jackson . core . JsonParser , com . fasterxml . jackson . databind . DeserializationContext ) { if ( jp instanceof org . mongojack . internal . JacksonDBCollectionProvider ) { K id = null ; java . lang . String collectionName = null ; com . fasterxml . jackson . core . JsonToken token = jp . getCurrentToken ( ) ; if ( token == ( com . fasterxml . jackson . core . JsonToken . VALUE_NULL ) ) { return null ; } if ( token == ( com . fasterxml . jackson . core . JsonToken . VALUE_EMBEDDED_OBJECT ) ) { java . lang . Object object = jp . getEmbeddedObject ( ) ; if ( object instanceof com . mongodb . DBRef ) { if ( ( keyDeserializer ) != null ) { id = keyDeserializer . deserialize ( jp , ctxt ) ; } else { id = ( ( K ) ( ( ( com . mongodb . DBRef ) ( object ) ) . getId ( ) ) ) ; } collectionName = ( ( com . mongodb . DBRef ) ( object ) ) . getCollectionName ( ) ; } else { throw ctxt . instantiationException ( org . mongojack . DBRef . class , ( "Don't<sp>know<sp>what<sp>to<sp>do<sp>with<sp>embedded<sp>object:<sp>" + object ) ) ; } } else if ( token == ( com . fasterxml . jackson . core . JsonToken . START_OBJECT ) ) { token = jp . nextValue ( ) ; while ( token != ( com . fasterxml . jackson . core . JsonToken . END_OBJECT ) ) { if ( jp . getCurrentName ( ) . equals ( "$id" ) ) { if ( ( keyDeserializer ) != null ) { id = keyDeserializer . deserialize ( jp , ctxt ) ; } else { id = ( ( K ) ( jp . getEmbeddedObject ( ) ) ) ; } } else if ( jp . getCurrentName ( ) . equals ( "$ref" ) ) { collectionName = jp . getText ( ) ; } else { } token = jp . nextValue ( ) ; } } if ( id == null ) { return null ; } if ( collectionName == null ) { throw ctxt . instantiationException ( org . mongojack . DBRef . class , "DBRef<sp>contains<sp>no<sp>collection<sp>name" ) ; } org . mongojack . JacksonDBCollection coll = ( ( org . mongojack . internal . JacksonDBCollectionProvider ) ( jp ) ) . getDBCollection ( ) ; org . mongojack . JacksonDBCollection < T , K > refColl = coll . getReferenceCollection ( collectionName , type , keyType ) ; return new org . mongojack . internal . FetchableDBRef < T , K > ( id , refColl ) ; } else { throw ctxt . instantiationException ( org . mongojack . DBRef . class , ( ( ( "DBRef<sp>can<sp>only<sp>be<sp>deserialised<sp>by<sp>this<sp>deserializer<sp>if<sp>parser<sp>implements<sp>" + ( org . mongojack . internal . JacksonDBCollectionProvider . class . getName ( ) ) ) + "<sp>parser<sp>is<sp>actually<sp>" ) + ( jp . getClass ( ) . getName ( ) ) ) ) ; } } | org . junit . Assert . assertNull ( deserializedDate ) |
NotificationEndPointUpdateCanSetNameAndAltId ( ) { java . lang . String expectedName = "newNotificationEndPointName" ; com . microsoft . windowsazure . services . media . entityoperations . EntityUpdateOperation updater = com . microsoft . windowsazure . services . media . models . NotificationEndPoint . update ( com . microsoft . windowsazure . services . media . models . NotificationEndPointEntityTest . sampleNotificationEndPointId ) . setName ( expectedName ) ; com . microsoft . windowsazure . services . media . implementation . content . NotificationEndPointType payload = ( ( com . microsoft . windowsazure . services . media . implementation . content . NotificationEndPointType ) ( updater . getRequestContents ( ) ) ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertEquals ( expectedName , payload . getName ( ) ) |
testMessageJobHasNoDueDateSet ( ) { configuration . setEnsureJobDueDateNotNull ( false ) ; runtimeService . startProcessInstanceByKey ( "simpleAsyncProcess" ) ; org . camunda . bpm . engine . runtime . Job job = managementService . createJobQuery ( ) . singleResult ( ) ; "<AssertPlaceHolder>" ; } getDuedate ( ) { return duedate ; } | org . junit . Assert . assertNull ( job . getDuedate ( ) ) |
apiFilterAuthenticatesDefaultUser_SetsApiUserInRequestIfValidUri ( ) { apiFilter . init ( mf . getMockFilterConfig ( ) ) ; when ( mf . getMockServletRequest ( ) . getRequestURI ( ) ) . thenReturn ( gov . nysenate . sage . filter . ApiFilterTest . validUri ) ; when ( mf . getMockServletRequest ( ) . getRemoteAddr ( ) ) . thenReturn ( "127.0.0.1" ) ; "<AssertPlaceHolder>" ; verify ( mf . getMockFilterChain ( ) , only ( ) ) . doFilter ( isA ( javax . servlet . ServletRequest . class ) , isA ( javax . servlet . ServletResponse . class ) ) ; } getMockServletRequest ( ) { return mockServletRequest ; } | org . junit . Assert . assertEquals ( "127.0.0.1" , mf . getMockServletRequest ( ) . getRemoteAddr ( ) ) |
delete_shouldDeleteEmptyDirectory ( ) { java . io . File file = new java . io . File ( "delete_shouldDeleteEmptyDirectory" ) ; file . mkdir ( ) ; com . silentcircle . silenttext . util . IOUtils . delete ( file ) ; "<AssertPlaceHolder>" ; } exists ( ) { if ( pendingRemoval ) { clear ( ) ; } return ( ! ( pendingRemoval ) ) && ( root . exists ( ) ) ; } | org . junit . Assert . assertFalse ( file . exists ( ) ) |
testEquals3 ( ) { s1 = new org . eclipse . tracecompass . ctf . core . trace . CTFStreamInput ( new org . eclipse . tracecompass . internal . ctf . core . trace . CTFStream ( org . eclipse . tracecompass . ctf . core . tests . shared . CtfTestTraceUtils . getTrace ( org . eclipse . tracecompass . ctf . core . tests . trace . CTFStreamInputTest . testTrace ) ) , org . eclipse . tracecompass . ctf . core . tests . trace . CTFStreamInputTest . createFile ( ) ) ; "<AssertPlaceHolder>" ; } createFile ( ) { java . io . File path = new java . io . File ( org . eclipse . tracecompass . ctf . core . tests . shared . CtfTestTraceUtils . getTrace ( org . eclipse . tracecompass . ctf . core . tests . trace . CTFStreamInputTest . testTrace ) . getPath ( ) ) ; final java . io . File [ ] listFiles = path . listFiles ( new java . io . FilenameFilter ( ) { @ org . eclipse . tracecompass . ctf . core . tests . trace . Override public boolean accept ( java . io . File dir , java . lang . String name ) { if ( name . contains ( "hann" ) ) { return true ; } return false ; } } ) ; org . junit . Assert . assertNotNull ( listFiles ) ; final java . io . File returnFile = listFiles [ 0 ] ; org . junit . Assert . assertNotNull ( returnFile ) ; return returnFile ; } | org . junit . Assert . assertEquals ( s1 , s1 ) |
testFindGroupByInstanceIndex ( ) { org . springframework . data . neo4j . support . Group group = new org . springframework . data . neo4j . support . SubGroup ( ) . persist ( ) ; group . setIndexLevelName ( "indexLevelNameValue" ) ; org . neo4j . graphdb . index . Index < org . springframework . data . neo4j . support . Node > subGroupIndex = graphDatabaseContext . getIndex ( org . springframework . data . neo4j . support . SubGroup . class ) ; final org . springframework . data . neo4j . support . Node found = subGroupIndex . get ( "indexLevelName" , "indexLevelNameValue" ) . getSingle ( ) ; final org . springframework . data . neo4j . support . SubGroup foundEntity = graphDatabaseContext . createEntityFromState ( found , org . springframework . data . neo4j . support . SubGroup . class ) ; "<AssertPlaceHolder>" ; } createEntityFromState ( S extends org . springframework . data . neo4j . support . PropertyContainer , java . lang . Class ) { if ( state == null ) throw new java . lang . IllegalArgumentException ( "state<sp>has<sp>to<sp>be<sp>either<sp>a<sp>Node<sp>or<sp>Relationship,<sp>not<sp>null" ) ; return getTypeRepresentationStrategy ( state , type ) . createEntity ( state , type ) ; } | org . junit . Assert . assertEquals ( group , foundEntity ) |
testCreateExistingChildCategory ( ) { org . pentaho . marketplace . domain . model . factories . interfaces . ICategoryFactory factory = this . createFactory ( ) ; java . lang . String parentName = "ParentCategory" ; java . lang . String childName = "ChildCategory" ; org . pentaho . marketplace . domain . model . entities . interfaces . ICategory parent = factory . create ( parentName ) ; org . pentaho . marketplace . domain . model . entities . interfaces . ICategory expectedChild = factory . create ( childName , parent ) ; org . pentaho . marketplace . domain . model . entities . interfaces . ICategory actualChild = factory . create ( childName , parent ) ; "<AssertPlaceHolder>" ; } create ( java . lang . String , java . lang . String ) { org . pentaho . marketplace . domain . model . entities . interfaces . IDomainStatusMessage domainStatusMessage = new org . pentaho . marketplace . domain . model . entities . DomainStatusMessage ( ) ; domainStatusMessage . setCode ( code ) ; domainStatusMessage . setMessage ( message ) ; return domainStatusMessage ; } | org . junit . Assert . assertSame ( expectedChild , actualChild ) |
testUpdateCharObject_Null ( ) { com . jmethods . catatumbo . entities . CharObject entity = new com . jmethods . catatumbo . entities . CharObject ( ) ; entity . setSex ( 'M' ) ; entity = com . jmethods . catatumbo . EntityManagerTest . em . insert ( entity ) ; entity . setSex ( null ) ; entity = com . jmethods . catatumbo . EntityManagerTest . em . update ( entity ) ; entity = com . jmethods . catatumbo . EntityManagerTest . em . load ( com . jmethods . catatumbo . entities . CharObject . class , entity . getId ( ) ) ; "<AssertPlaceHolder>" ; } getSex ( ) { return sex ; } | org . junit . Assert . assertTrue ( ( ( entity . getSex ( ) ) == null ) ) |
formatNumerals ( ) { zemberek . morphology . TurkishMorphology morphology = zemberek . morphology . TurkishMorphology . builder ( ) . disableCache ( ) . build ( ) ; java . lang . String [ ] inputs = new java . lang . String [ ] { "1e" , "4ten" , "123" , "12,5ten" } ; java . lang . String [ ] expected = new java . lang . String [ ] { "1'e" , "4'ten" , "123'" , "12,5ten" } ; zemberek . morphology . analysis . WordAnalysisSurfaceFormatter formatter = new zemberek . morphology . analysis . WordAnalysisSurfaceFormatter ( ) ; int i = 0 ; for ( java . lang . String input : inputs ) { zemberek . morphology . analysis . WordAnalysis results = morphology . analyze ( input ) ; for ( zemberek . morphology . analysis . SingleAnalysis result : results ) { if ( ( result . getDictionaryItem ( ) . primaryPos ) == ( zemberek . core . turkish . PrimaryPos . Numeral ) ) { "<AssertPlaceHolder>" ; } } i ++ ; } } format ( zemberek . morphology . analysis . SingleAnalysis , java . lang . String ) { zemberek . morphology . lexicon . DictionaryItem item = analysis . getDictionaryItem ( ) ; java . lang . String ending = analysis . getEnding ( ) ; if ( ( apostrophe != null ) || ( apostropheRequired ( analysis ) ) ) { if ( apostrophe == null ) { apostrophe = "'" ; } return ( ending . length ( ) ) > 0 ? ( ( item . normalizedLemma ( ) ) + apostrophe ) + ending : item . normalizedLemma ( ) ; } else { if ( item . attributes . contains ( RootAttribute . NoQuote ) ) { return ( item . normalizedLemma ( ) ) + ending ; } else { return ( analysis . getStem ( ) ) + ending ; } } } | org . junit . Assert . assertEquals ( expected [ i ] , formatter . format ( result , "'" ) ) |
testSmallPlaintextWithRawKey ( ) { byte [ ] keyValue = com . google . crypto . tink . subtle . Random . randBytes ( com . google . crypto . tink . mac . MacFactoryTest . HMAC_KEY_SIZE ) ; com . google . crypto . tink . proto . Keyset . Key primary = com . google . crypto . tink . TestUtil . createKey ( com . google . crypto . tink . TestUtil . createHmacKeyData ( keyValue , 16 ) , 42 , KeyStatusType . ENABLED , OutputPrefixType . RAW ) ; com . google . crypto . tink . KeysetHandle keysetHandle = com . google . crypto . tink . TestUtil . createKeysetHandle ( com . google . crypto . tink . TestUtil . createKeyset ( primary ) ) ; com . google . crypto . tink . Mac mac = com . google . crypto . tink . mac . MacFactory . getPrimitive ( keysetHandle ) ; byte [ ] plaintext = "blah" . getBytes ( "UTF-8" ) ; byte [ ] tag = mac . computeMac ( plaintext ) ; "<AssertPlaceHolder>" ; try { mac . verifyMac ( tag , plaintext ) ; } catch ( java . security . GeneralSecurityException e ) { org . junit . Assert . fail ( "Valid<sp>MAC,<sp>should<sp>not<sp>throw<sp>exception" ) ; } } computeMac ( byte [ ] ) { if ( primitives . getPrimary ( ) . getOutputPrefixType ( ) . equals ( OutputPrefixType . LEGACY ) ) { return com . google . crypto . tink . subtle . Bytes . concat ( primitives . getPrimary ( ) . getIdentifier ( ) , primitives . getPrimary ( ) . getPrimitive ( ) . computeMac ( com . google . crypto . tink . subtle . Bytes . concat ( data , formatVersion ) ) ) ; } return com . google . crypto . tink . subtle . Bytes . concat ( primitives . getPrimary ( ) . getIdentifier ( ) , primitives . getPrimary ( ) . getPrimitive ( ) . computeMac ( data ) ) ; } | org . junit . Assert . assertEquals ( 16 , tag . length ) |
testValidMinCardinality ( ) { com . github . anno4j . Anno4j anno4j = new com . github . anno4j . Anno4j ( ) ; com . github . anno4j . ValidatedTransaction transaction = anno4j . createValidatedTransaction ( ) ; transaction . begin ( ) ; com . github . anno4j . transaction . ValidatedTransactionTest . SpecialTestResource resource = transaction . createObject ( com . github . anno4j . transaction . ValidatedTransactionTest . SpecialTestResource . class ) ; resource . setCardinality ( com . google . common . collect . Sets . newHashSet ( 1 , 2 ) ) ; resource . setQualifiedCardinality ( com . google . common . collect . Sets . < com . github . anno4j . transaction . ValidatedTransactionTest . TestResource > newHashSet ( resource ) ) ; boolean exceptionThrown = false ; try { transaction . commit ( ) ; } catch ( com . github . anno4j . ValidatedTransaction e ) { exceptionThrown = true ; } "<AssertPlaceHolder>" ; } commit ( ) { connection . commit ( ) ; } | org . junit . Assert . assertFalse ( exceptionThrown ) |
verifyThatDataIsAvailableInOrcNocomp ( ) { java . util . List < java . lang . String > expected = java . util . Arrays . asList ( "A\tB" , "C\tD" , "E\tF" ) ; java . util . List < java . lang . String > actual = hiveShell . executeQuery ( "select<sp>*<sp>from<sp>foo_orc_nocomp" ) ; "<AssertPlaceHolder>" ; } executeQuery ( java . nio . file . Path ) { return executeQuery ( java . nio . charset . Charset . defaultCharset ( ) , script ) ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testClosedClosedDescending ( ) { org . apache . commons . functor . range . DoubleRange range = org . apache . commons . functor . range . Ranges . doubleRange ( 5.0 , BoundType . CLOSED , ( - 5.0 ) , BoundType . CLOSED , ( - 3.0 ) ) ; java . util . List < java . lang . Double > expected = java . util . Arrays . asList ( 5.0 , 2.0 , ( - 1.0 ) , ( - 4.0 ) ) ; java . util . Collection < java . lang . Double > elements = org . apache . commons . functor . generator . loop . IteratorToGeneratorAdapter . adapt ( range ) . toCollection ( ) ; "<AssertPlaceHolder>" ; } toCollection ( ) { return new org . apache . commons . functor . generator . util . CollectionTransformer < E , java . util . Collection < E > > ( new java . util . ArrayList < E > ( ) ) ; } | org . junit . Assert . assertEquals ( expected , elements ) |
testNewInstance_noArgsConstructorPresent ( ) { java . util . ArrayList < ? > i = org . objenesis . instantiator . util . ClassUtils . newInstance ( java . util . ArrayList . class ) ; "<AssertPlaceHolder>" ; } newInstance ( java . lang . Class ) { try { return clazz . newInstance ( ) ; } catch ( java . lang . InstantiationException | java . lang . IllegalAccessException e ) { throw new org . objenesis . ObjenesisException ( e ) ; } } | org . junit . Assert . assertTrue ( i . isEmpty ( ) ) |
alternativeCoveragePlanShouldFailIfNoPrimaryPartitionAvailable ( ) { final com . basho . riak . client . api . RiakClient client = new com . basho . riak . client . api . RiakClient ( cluster ) ; final com . basho . riak . client . core . query . Namespace ns = new com . basho . riak . client . core . query . Namespace ( com . basho . riak . client . core . query . Namespace . DEFAULT_BUCKET_TYPE , bucketName . toString ( ) ) ; final com . basho . riak . client . api . commands . buckets . FetchBucketProperties fetchProps = new com . basho . riak . client . api . commands . buckets . FetchBucketProperties . Builder ( ns ) . build ( ) ; final com . basho . riak . client . core . operations . FetchBucketPropsOperation . Response fetchResponse = client . execute ( fetchProps ) ; final com . basho . riak . client . core . query . BucketProperties bp = fetchResponse . getBucketProperties ( ) ; final java . lang . Integer nVal = bp . getNVal ( ) ; final int minPartitions = 5 ; final com . basho . riak . client . api . commands . kv . CoveragePlan cmd = CoveragePlan . Builder . create ( defaultNamespace ( ) ) . withMinPartitions ( minPartitions ) . build ( ) ; final com . basho . riak . client . api . commands . kv . CoveragePlan . Response response = client . execute ( cmd ) ; final com . basho . riak . client . core . operations . itest . List < com . basho . riak . client . core . operations . CoveragePlanOperation . Response . CoverageEntry > coverageEntries = new com . basho . riak . client . core . operations . itest . LinkedList ( ) ; for ( com . basho . riak . client . core . operations . CoveragePlanOperation . Response . CoverageEntry ce : response ) { coverageEntries . add ( ce ) ; } com . basho . riak . client . core . operations . CoveragePlanOperation . Response . CoverageEntry failedEntry = coverageEntries . get ( 0 ) ; com . basho . riak . client . core . operations . itest . List < com . basho . riak . client . core . operations . CoveragePlanOperation . Response . CoverageEntry > unavailableCoverageEntries = new com . basho . riak . client . core . operations . itest . LinkedList ( ) ; for ( int i = 0 ; i < ( nVal - 1 ) ; i ++ ) { unavailableCoverageEntries . add ( failedEntry ) ; final com . basho . riak . client . api . commands . kv . CoveragePlan cmdAlternative = CoveragePlan . Builder . create ( defaultNamespace ( ) ) . withMinPartitions ( minPartitions ) . withReplaceCoverageEntry ( failedEntry ) . withUnavailableCoverageEntries ( unavailableCoverageEntries ) . build ( ) ; final com . basho . riak . client . api . commands . kv . CoveragePlan . Response responseAlternative = client . execute ( cmdAlternative ) ; "<AssertPlaceHolder>" ; failedEntry = responseAlternative . iterator ( ) . next ( ) ; } unavailableCoverageEntries . add ( failedEntry ) ; final com . basho . riak . client . api . commands . kv . CoveragePlan cmdAlternativeFailing = CoveragePlan . Builder . create ( defaultNamespace ( ) ) . withMinPartitions ( minPartitions ) . withReplaceCoverageEntry ( failedEntry ) . withUnavailableCoverageEntries ( unavailableCoverageEntries ) . build ( ) ; exception . expect ( java . util . concurrent . ExecutionException . class ) ; exception . expectMessage ( "com.basho.riak.client.core.netty.RiakResponseException:<sp>primary_partition_unavailable" ) ; client . execute ( cmdAlternativeFailing ) ; } iterator ( ) { if ( isStreaming ( ) ) { assert ( chunkedResponseIterator ) != null ; return chunkedResponseIterator ; } throw new java . lang . UnsupportedOperationException ( "Iterating<sp>is<sp>only<sp>supported<sp>for<sp>streamable<sp>response" ) ; } | org . junit . Assert . assertTrue ( responseAlternative . iterator ( ) . hasNext ( ) ) |
testSetInboundStandardMode ( ) { gov . hhs . fha . nhinc . configuration . jmx . PassthruMXBeanRegistry registry = gov . hhs . fha . nhinc . configuration . jmx . PassthruMXBeanRegistry . getInstance ( ) ; gov . hhs . fha . nhinc . configuration . IConfiguration . serviceEnum serviceName = gov . hhs . fha . nhinc . configuration . IConfiguration . serviceEnum . QueryForDocuments ; gov . hhs . fha . nhinc . configuration . IConfiguration . directionEnum direction = gov . hhs . fha . nhinc . configuration . IConfiguration . directionEnum . Inbound ; boolean status = true ; gov . hhs . fha . nhinc . docquery . configuration . jmx . DocumentQuery20WebServices docquery20 = mock ( gov . hhs . fha . nhinc . docquery . configuration . jmx . DocumentQuery20WebServices . class ) ; when ( docquery20 . getServiceName ( ) ) . thenReturn ( serviceEnum . QueryForDocuments ) ; when ( docquery20 . isInboundStandard ( ) ) . thenReturn ( status ) ; registry . registerWebServiceMXBean ( docquery20 ) ; boolean standard = registry . isStandard ( serviceName , direction ) ; "<AssertPlaceHolder>" ; } isStandard ( gov . hhs . fha . nhinc . configuration . IConfiguration . serviceEnum , gov . hhs . fha . nhinc . configuration . IConfiguration . directionEnum ) { boolean standardMode = false ; for ( gov . hhs . fha . nhinc . configuration . jmx . WebServicesMXBean b : registeredBeans ) { if ( ( ( gov . hhs . fha . nhinc . configuration . jmx . PassthruMXBeanRegistry . isOutbound ( direction ) ) && ( b . getServiceName ( ) . equals ( serviceName ) ) ) && ( b . isOutboundStandard ( ) ) ) { standardMode = true ; } if ( ( ( gov . hhs . fha . nhinc . configuration . jmx . PassthruMXBeanRegistry . isInbound ( direction ) ) && ( b . getServiceName ( ) . equals ( serviceName ) ) ) && ( b . isInboundStandard ( ) ) ) { standardMode = true ; } } return standardMode ; } | org . junit . Assert . assertEquals ( true , standard ) |
testRemoveIsUnsupported ( ) { org . gedcomx . records . RecordSet recordSet1 = new org . gedcomx . records . RecordSet ( ) ; recordSet1 . setMetadata ( org . gedcomx . util . TestRecordSetWriter . getMetadataFromFile ( ) ) ; java . util . List < org . gedcomx . Gedcomx > records = org . gedcomx . util . JsonRecordSetIteratorTest . getRecordsFromRecordSetFile ( ) ; recordSet1 . setRecords ( records ) ; java . io . ByteArrayOutputStream bos = new java . io . ByteArrayOutputStream ( ) ; com . fasterxml . jackson . databind . ObjectMapper objectMapper = org . gedcomx . rt . json . GedcomJacksonModule . createObjectMapper ( ) ; objectMapper . writeValue ( bos , recordSet1 ) ; org . gedcomx . records . RecordSet recordSet2 = new org . gedcomx . records . RecordSet ( ) ; recordSet2 . setRecords ( new java . util . ArrayList < org . gedcomx . Gedcomx > ( ) ) ; byte [ ] bytes = bos . toByteArray ( ) ; java . io . InputStream inputStream = new java . io . ByteArrayInputStream ( bytes ) ; org . gedcomx . util . JsonRecordSetIterator jsonRecordSetIterator = new org . gedcomx . util . JsonRecordSetIterator ( inputStream , false ) ; try { jsonRecordSetIterator . remove ( ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . UnsupportedOperationException e ) { } jsonRecordSetIterator . close ( ) ; } remove ( ) { throw new java . lang . UnsupportedOperationException ( ) ; } | org . junit . Assert . assertTrue ( false ) |
getBaseCalendarsAndNotResourceCalendars ( ) { java . util . List < org . libreplan . business . calendars . entities . BaseCalendar > baseCalendars = baseCalendarDAO . getBaseCalendars ( ) ; int previous = baseCalendars . size ( ) ; org . libreplan . business . calendars . entities . BaseCalendar calendar1 = org . libreplan . business . test . calendars . entities . BaseCalendarTest . createBasicCalendar ( ) ; calendar1 . setName ( "Test1" ) ; org . libreplan . business . calendars . entities . BaseCalendar calendar2 = org . libreplan . business . test . calendars . entities . BaseCalendarTest . createBasicCalendar ( ) ; calendar1 . setName ( "Test2" ) ; org . libreplan . business . resources . entities . Worker worker = org . libreplan . business . test . resources . daos . ResourceDAOTest . givenValidWorker ( ) ; org . libreplan . business . calendars . entities . ResourceCalendar resourceCalendar = org . libreplan . business . calendars . entities . ResourceCalendar . create ( ) ; resourceCalendar . setName ( "testResourceCalendar" ) ; org . libreplan . business . test . calendars . entities . BaseCalendarTest . setHoursForAllDays ( resourceCalendar , 8 ) ; worker . setCalendar ( resourceCalendar ) ; baseCalendarDAO . save ( calendar1 ) ; baseCalendarDAO . save ( calendar2 ) ; resourceDAO . save ( worker ) ; baseCalendarDAO . flush ( ) ; resourceDAO . flush ( ) ; baseCalendars = baseCalendarDAO . getBaseCalendars ( ) ; "<AssertPlaceHolder>" ; } equalTo ( T extends java . lang . Comparable ) { return org . zkoss . ganttz . data . constraint . ConstraintOnComparableValues . instantiate ( org . zkoss . ganttz . data . constraint . ConstraintOnComparableValues . ComparisonType . EQUAL_TO , value ) ; } | org . junit . Assert . assertThat ( baseCalendars . size ( ) , org . hamcrest . CoreMatchers . equalTo ( ( previous + 2 ) ) ) |
shouldCheckGetBooksURI ( ) { javax . ws . rs . core . Response response = org . agoncal . book . javaee7 . chapter15 . ex05 . ItemRestService05IT . client . target ( "http://localhost:8282/05/items/books" ) . request ( ) . get ( ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( 200 , response . getStatus ( ) ) |
testGetAll ( ) { jp . primecloud . auto . zabbix . model . trigger . TriggerGetParam param = new jp . primecloud . auto . zabbix . model . trigger . TriggerGetParam ( ) ; param . setHostids ( java . util . Arrays . asList ( "10001" ) ) ; param . setOutput ( "extend" ) ; java . util . List < jp . primecloud . auto . zabbix . model . trigger . Trigger > triggers = client . trigger ( ) . get ( param ) ; for ( jp . primecloud . auto . zabbix . model . trigger . Trigger trigger : triggers ) { log . trace ( org . apache . commons . lang . builder . ReflectionToStringBuilder . toString ( trigger , ToStringStyle . SHORT_PREFIX_STYLE ) ) ; } "<AssertPlaceHolder>" ; } get ( java . lang . String ) { if ( ( jp . primecloud . auto . ui . util . ViewProperties . userBundle ) != null ) { try { return jp . primecloud . auto . ui . util . ViewProperties . userBundle . getString ( key ) ; } catch ( java . util . MissingResourceException ignore ) { } } try { return jp . primecloud . auto . ui . util . ViewProperties . defaultBundle . getString ( key ) ; } catch ( java . util . MissingResourceException ignore ) { } return "" ; } | org . junit . Assert . assertTrue ( ( ( triggers . size ( ) ) > 0 ) ) |
filter ( ) { java . util . List < java . lang . Object > objects = new java . util . ArrayList < java . lang . Object > ( ) ; objects . add ( comparable ) ; objects . add ( comparable ) ; doReturn ( 0 ) . doReturn ( 1 ) . when ( comparable ) . compareTo ( any ( ) ) ; doReturn ( comparable ) . when ( etFilter ) . getAttributeValue ( comparable ) ; java . util . List < java . lang . Object > actualObjects = etFilter . filter ( objects ) ; "<AssertPlaceHolder>" ; } get ( byte [ ] ) { return getClient ( ) . get ( key ) ; } | org . junit . Assert . assertEquals ( comparable , actualObjects . get ( 0 ) ) |
testNullInitialization ( ) { timeFrameEditor . init ( org . dashbuilder . displayer . client . widgets . filter . TimeFrameEditorTest . UNDEFINED , changeCommand ) ; "<AssertPlaceHolder>" ; verify ( timeFrameView ) . init ( timeFrameEditor ) ; verify ( timeFrameView ) . clearFirstMonthSelector ( ) ; verify ( timeFrameView , times ( org . dashbuilder . dataset . date . Month . values ( ) . length ) ) . addFirstMonthItem ( any ( org . dashbuilder . dataset . date . Month . class ) ) ; verify ( timeFrameView ) . setSelectedFirstMonthIndex ( ( ( Month . JANUARY . getIndex ( ) ) - 1 ) ) ; } init ( java . io . ByteArrayOutputStream , java . lang . String ) { this . originalPath = ioService . get ( new java . net . URI ( uri ) ) ; this . zipWriter = new org . guvnor . common . services . backend . archive . ZipWriter ( outputStream ) ; } | org . junit . Assert . assertEquals ( timeFrameView , timeFrameEditor . view ) |
testLockAdministratieveHandelingTweeKeerMetDaardoorExceptie ( ) { final boolean resultaat = administratieveHandelingLockRepository . vergrendelAlsNogNietIsVerwerkt ( nl . bzk . brp . levering . dataaccess . repository . lezenschrijven . AdministratieveHandelingVergrendelRepositoryIntegratieTest . ADMINISTRATIEVE_HANDELING_ID ) ; "<AssertPlaceHolder>" ; tryLockFromSql ( ) ; } vergrendelAlsNogNietIsVerwerkt ( java . lang . Long ) { if ( isNietVerwerktEnVergrendelingGeplaatst ( administratieveHandelingId ) ) { return true ; } return false ; } | org . junit . Assert . assertTrue ( resultaat ) |
testSearchCountByLanguage ( ) { java . lang . String velocityLanguage = com . liferay . portal . kernel . template . TemplateConstants . LANG_TYPE_VM ; addTemplate ( com . liferay . portal . kernel . test . util . RandomTestUtil . randomLong ( ) , com . liferay . portal . kernel . test . util . RandomTestUtil . randomLong ( ) , com . liferay . portal . kernel . util . StringUtil . randomString ( ) , StringPool . BLANK , StringPool . BLANK , velocityLanguage , getTestTemplateScript ( velocityLanguage ) , WorkflowConstants . STATUS_APPROVED ) ; addTemplate ( com . liferay . portal . kernel . test . util . RandomTestUtil . randomLong ( ) , com . liferay . portal . kernel . test . util . RandomTestUtil . randomLong ( ) , com . liferay . portal . kernel . util . StringUtil . randomString ( ) , StringPool . BLANK , StringPool . BLANK , velocityLanguage , getTestTemplateScript ( velocityLanguage ) , WorkflowConstants . STATUS_APPROVED ) ; java . lang . String freeMarkerLanguage = com . liferay . portal . kernel . template . TemplateConstants . LANG_TYPE_FTL ; addTemplate ( com . liferay . portal . kernel . test . util . RandomTestUtil . randomLong ( ) , com . liferay . portal . kernel . test . util . RandomTestUtil . randomLong ( ) , com . liferay . portal . kernel . util . StringUtil . randomString ( ) , StringPool . BLANK , StringPool . BLANK , freeMarkerLanguage , getTestTemplateScript ( freeMarkerLanguage ) , WorkflowConstants . STATUS_APPROVED ) ; int count = com . liferay . dynamic . data . mapping . service . DDMTemplateLocalServiceUtil . searchCount ( com . liferay . portal . kernel . test . util . TestPropsValues . getCompanyId ( ) , new long [ ] { group . getGroupId ( ) } , null , null , com . liferay . dynamic . data . mapping . service . test . DDMTemplateLocalServiceTest . _resourceClassNameId , null , null , null , null , TemplateConstants . LANG_TYPE_VM , WorkflowConstants . STATUS_APPROVED , true ) ; "<AssertPlaceHolder>" ; } getGroupId ( ) { return _groupId ; } | org . junit . Assert . assertEquals ( 2 , count ) |
testReshuffleAfterSlidingWindows ( ) { org . apache . beam . sdk . values . PCollection < org . apache . beam . sdk . values . KV < java . lang . String , java . lang . Integer > > input = pipeline . apply ( org . apache . beam . sdk . transforms . Create . of ( org . apache . beam . sdk . transforms . ReshuffleTest . ARBITRARY_KVS ) . withCoder ( org . apache . beam . sdk . coders . KvCoder . of ( org . apache . beam . sdk . coders . StringUtf8Coder . of ( ) , org . apache . beam . sdk . coders . VarIntCoder . of ( ) ) ) ) . apply ( org . apache . beam . sdk . transforms . windowing . Window . into ( org . apache . beam . sdk . transforms . windowing . FixedWindows . of ( org . joda . time . Duration . standardMinutes ( 10L ) ) ) ) ; org . apache . beam . sdk . values . PCollection < org . apache . beam . sdk . values . KV < java . lang . String , java . lang . Integer > > output = input . apply ( org . apache . beam . sdk . transforms . Reshuffle . of ( ) ) ; org . apache . beam . sdk . testing . PAssert . that ( output ) . containsInAnyOrder ( org . apache . beam . sdk . transforms . ReshuffleTest . ARBITRARY_KVS ) ; "<AssertPlaceHolder>" ; pipeline . run ( ) ; } getWindowingStrategy ( ) { return windowingStrategy ; } | org . junit . Assert . assertEquals ( input . getWindowingStrategy ( ) , output . getWindowingStrategy ( ) ) |
testSaveCriterionType ( ) { org . libreplan . business . resources . entities . CriterionType criterionType = org . libreplan . business . test . resources . daos . CriterionTypeDAOTest . createValidCriterionType ( ) ; criterionTypeDAO . save ( criterionType ) ; "<AssertPlaceHolder>" ; } exists ( PK extends java . io . Serializable ) { return ( getSession ( ) . createCriteria ( entityClass ) . add ( org . hibernate . criterion . Restrictions . idEq ( id ) ) . setProjection ( org . hibernate . criterion . Projections . id ( ) ) . uniqueResult ( ) ) != null ; } | org . junit . Assert . assertTrue ( criterionTypeDAO . exists ( criterionType . getId ( ) ) ) |
testExeResponseLiteralDataType ( ) { if ( org . geotools . data . wps . ParsingTest . DISABLE ) { return ; } java . io . File file = org . geotools . TestData . file ( this , "LiteralDataTypeTestFile.xml" ) ; java . io . BufferedReader in = new java . io . BufferedReader ( new java . io . FileReader ( file ) ) ; org . geotools . xsd . Configuration config = new org . geotools . wps . WPSConfiguration ( ) ; org . geotools . xsd . Parser parser = new org . geotools . xsd . Parser ( config ) ; java . lang . Object object = parser . parse ( in ) ; net . opengis . wps10 . ExecuteResponseType exeResponse = null ; if ( object instanceof net . opengis . wps10 . ExecuteResponseType ) { exeResponse = ( ( net . opengis . wps10 . ExecuteResponseType ) ( object ) ) ; } net . opengis . wps10 . OutputDataType odt = ( ( net . opengis . wps10 . OutputDataType ) ( exeResponse . getProcessOutputs ( ) . getOutput ( ) . get ( 0 ) ) ) ; java . lang . String dataType = odt . getData ( ) . getLiteralData ( ) . getDataType ( ) ; "<AssertPlaceHolder>" ; } getDataType ( ) { return dataType ; } | org . junit . Assert . assertNotNull ( dataType ) |
testWithEmptyHsqldb ( ) { org . qcri . rheem . core . api . Configuration configuration = new org . qcri . rheem . core . api . Configuration ( ) ; org . qcri . rheem . core . api . Job job = mock ( org . qcri . rheem . core . api . Job . class ) ; when ( job . getConfiguration ( ) ) . thenReturn ( configuration ) ; org . qcri . rheem . core . platform . CrossPlatformExecutor cpe = new org . qcri . rheem . core . platform . CrossPlatformExecutor ( job , new org . qcri . rheem . core . profiling . FullInstrumentationStrategy ( ) ) ; when ( job . getCrossPlatformExecutor ( ) ) . thenReturn ( cpe ) ; final org . qcri . rheem . java . execution . JavaExecutor javaExecutor = new org . qcri . rheem . java . execution . JavaExecutor ( org . qcri . rheem . java . platform . JavaPlatform . getInstance ( ) , job ) ; org . qcri . rheem . jdbc . test . HsqldbPlatform hsqldbPlatform = new org . qcri . rheem . jdbc . test . HsqldbPlatform ( ) ; try ( java . sql . Connection jdbcConnection = hsqldbPlatform . createDatabaseDescriptor ( configuration ) . createJdbcConnection ( ) ) { final java . sql . Statement statement = jdbcConnection . createStatement ( ) ; statement . execute ( "CREATE<sp>TABLE<sp>testWithEmptyHsqldb<sp>(a<sp>INT,<sp>b<sp>VARCHAR(6));" ) ; } final org . qcri . rheem . core . plan . rheemplan . ExecutionOperator filterOperator = new org . qcri . rheem . jdbc . test . HsqldbFilterOperator ( new org . qcri . rheem . core . function . PredicateDescriptor ( ( x ) -> false , org . qcri . rheem . basic . data . Record . class ) ) ; final org . qcri . rheem . jdbc . channels . SqlQueryChannel sqlQueryChannel = new org . qcri . rheem . jdbc . channels . SqlQueryChannel ( org . qcri . rheem . jdbc . test . HsqldbPlatform . getInstance ( ) . getSqlQueryChannelDescriptor ( ) , filterOperator . getOutput ( 0 ) ) ; org . qcri . rheem . jdbc . channels . SqlQueryChannel . Instance sqlQueryChannelInstance = sqlQueryChannel . createInstance ( hsqldbPlatform . createExecutor ( job ) , mock ( OptimizationContext . OperatorContext . class ) , 0 ) ; sqlQueryChannelInstance . setSqlQuery ( "SELECT<sp>*<sp>FROM<sp>testWithEmptyHsqldb;" ) ; org . qcri . rheem . core . plan . executionplan . ExecutionTask producer = new org . qcri . rheem . core . plan . executionplan . ExecutionTask ( filterOperator ) ; producer . setOutputChannel ( 0 , sqlQueryChannel ) ; org . qcri . rheem . java . channels . StreamChannel . Instance streamChannelInstance = new org . qcri . rheem . java . channels . StreamChannel ( org . qcri . rheem . java . channels . StreamChannel . DESCRIPTOR , mock ( org . qcri . rheem . core . plan . rheemplan . OutputSlot . class ) ) . createInstance ( javaExecutor , mock ( OptimizationContext . OperatorContext . class ) , 0 ) ; org . qcri . rheem . jdbc . operators . SqlToStreamOperator sqlToStreamOperator = new org . qcri . rheem . jdbc . operators . SqlToStreamOperator ( org . qcri . rheem . jdbc . test . HsqldbPlatform . getInstance ( ) ) ; evaluate ( sqlToStreamOperator , new org . qcri . rheem . core . platform . ChannelInstance [ ] { sqlQueryChannelInstance } , new org . qcri . rheem . core . platform . ChannelInstance [ ] { streamChannelInstance } ) ; java . util . List < org . qcri . rheem . basic . data . Record > output = streamChannelInstance . < org . qcri . rheem . basic . data . Record > provideStream ( ) . collect ( java . util . stream . Collectors . toList ( ) ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return this . counts . isEmpty ( ) ; } | org . junit . Assert . assertTrue ( output . isEmpty ( ) ) |
testJBRULES_2995 ( ) { final java . lang . String str = "package<sp>org.drools.compiler\n" + ( ( ( ( ( ( "rule<sp>r1\n" + "when\n" ) + "<sp>Primitives(<sp>classAttr<sp>==<sp>java.lang.String.class,<sp>\n" ) + "<sp>eval(classAttr.equals(<sp>java.lang.String.class<sp>)<sp>),\n" ) + "<sp>classAttr<sp>==<sp>String.class<sp>)\n" ) + "then\n" ) + "end\n" ) ; final org . kie . api . KieBase kbase = loadKnowledgeBaseFromString ( str ) ; final org . kie . api . runtime . KieSession ksession = createKnowledgeSession ( kbase ) ; final org . drools . compiler . Primitives primitives = new org . drools . compiler . Primitives ( ) ; primitives . setClassAttr ( java . lang . String . class ) ; ksession . insert ( primitives ) ; final int rules = ksession . fireAllRules ( ) ; "<AssertPlaceHolder>" ; } fireAllRules ( ) { return 0 ; } | org . junit . Assert . assertEquals ( 1 , rules ) |
parse_parens ( ) { com . psddev . dari . db . Predicate pred = parser . parse ( "a<sp>=<sp>1<sp>and<sp>(b<sp>=<sp>2<sp>or<sp>c<sp>=<sp>3)" ) ; com . psddev . dari . db . Predicate expect = com . psddev . dari . db . CompoundPredicate . combine ( PredicateParser . AND_OPERATOR , new com . psddev . dari . db . ComparisonPredicate ( PredicateParser . EQUALS_ANY_OPERATOR , false , "a" , java . util . Arrays . asList ( "1" ) ) , com . psddev . dari . db . CompoundPredicate . combine ( PredicateParser . OR_OPERATOR , new com . psddev . dari . db . ComparisonPredicate ( PredicateParser . EQUALS_ANY_OPERATOR , false , "b" , java . util . Arrays . asList ( "2" ) ) , new com . psddev . dari . db . ComparisonPredicate ( PredicateParser . EQUALS_ANY_OPERATOR , false , "c" , java . util . Arrays . asList ( "3" ) ) ) ) ; "<AssertPlaceHolder>" ; } parse ( byte [ ] ) { try { return parseAny ( bytes ) ; } catch ( org . codehaus . jackson . JsonParseException error ) { throw new com . psddev . dari . util . JsonParsingException ( "Can't<sp>parse<sp>JSON<sp>bytes!" , error ) ; } catch ( java . io . IOException error ) { throw new java . lang . IllegalStateException ( error ) ; } } | org . junit . Assert . assertEquals ( expect , pred ) |
testSchemaValidation ( ) { io . cdap . plugin . JSONParser . Config config = new io . cdap . plugin . JSONParser . Config ( "body" , "" , io . cdap . plugin . JSONParserTest . OUTPUT2 . toString ( ) ) ; io . cdap . cdap . etl . api . Transform < io . cdap . cdap . api . data . format . StructuredRecord , io . cdap . cdap . api . data . format . StructuredRecord > transform = new io . cdap . plugin . JSONParser ( config ) ; io . cdap . cdap . etl . mock . common . MockPipelineConfigurer mockPipelineConfigurer = new io . cdap . cdap . etl . mock . common . MockPipelineConfigurer ( io . cdap . plugin . JSONParserTest . INPUT1 ) ; transform . configurePipeline ( mockPipelineConfigurer ) ; "<AssertPlaceHolder>" ; } getOutputSchema ( ) { java . util . List < io . cdap . cdap . api . data . schema . Schema . Field > fields = new java . util . ArrayList ( ) ; java . lang . String [ ] mappings = fieldTypeMapping . split ( "," ) ; for ( java . lang . String mapping : mappings ) { java . lang . String [ ] params = mapping . split ( ":" ) ; java . lang . String fieldName = params [ 0 ] . trim ( ) ; if ( com . google . common . base . Strings . isNullOrEmpty ( fieldName ) ) { throw new java . lang . IllegalArgumentException ( "Field<sp>name<sp>cannot<sp>be<sp>null<sp>or<sp>empty." ) ; } else if ( ( ( params . length ) < 2 ) || ( com . google . common . base . Strings . isNullOrEmpty ( params [ 1 ] ) ) ) { throw new java . lang . IllegalArgumentException ( ( "Type<sp>cannot<sp>be<sp>null.<sp>Please<sp>specify<sp>type<sp>for<sp>" + fieldName ) ) ; } io . cdap . cdap . api . data . schema . Schema . Field field = Schema . Field . of ( fieldName , io . cdap . cdap . api . data . schema . Schema . nullableOf ( io . cdap . cdap . api . data . schema . Schema . of ( Schema . Type . valueOf ( params [ 1 ] . trim ( ) . toUpperCase ( ) ) ) ) ) ; if ( fields . contains ( field ) ) { throw new java . lang . IllegalArgumentException ( java . lang . String . format ( "Field<sp>%s<sp>already<sp>has<sp>type<sp>specified.<sp>Duplicate<sp>field<sp>%s" , fieldName , fieldName ) ) ; } else { fields . add ( field ) ; } } return io . cdap . cdap . api . data . schema . Schema . recordOf ( "record" , fields ) ; } | org . junit . Assert . assertEquals ( io . cdap . plugin . JSONParserTest . OUTPUT2 , mockPipelineConfigurer . getOutputSchema ( ) ) |
testDisableNegatedPatterns ( ) { page . disableNegatedPatterns ( ) ; "<AssertPlaceHolder>" ; } isNegatedPatternEnabled ( ) { return negatedPatternEnabled ; } | org . junit . Assert . assertFalse ( page . isNegatedPatternEnabled ( ) ) |
testGroupShell ( ) { org . apache . log4j . Logger . getRootLogger ( ) . setLevel ( org . apache . log4j . Level . DEBUG ) ; org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; conf . set ( CommonConfigurationKeys . HADOOP_SECURITY_GROUP_MAPPING , "org.apache.hadoop.security.ShellBasedUnixGroupsMapping" ) ; org . apache . hadoop . security . Groups groups = new org . apache . hadoop . security . Groups ( conf ) ; java . lang . String username = java . lang . System . getProperty ( "user.name" ) ; java . util . List < java . lang . String > groupList = groups . getGroups ( username ) ; org . apache . hadoop . security . TestGroupFallback . LOG . info ( ( ( username + "<sp>has<sp>GROUPS:<sp>" ) + ( groupList . toString ( ) ) ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return blocks . size ( ) ; } | org . junit . Assert . assertTrue ( ( ( groupList . size ( ) ) > 0 ) ) |
runTest ( ) { boolean result = checkNoError ( "Social_Profiles_Get_People_Managed" ) ; "<AssertPlaceHolder>" ; } getNoErrorMsg ( ) { return noErrorMsg ; } | org . junit . Assert . assertTrue ( getNoErrorMsg ( ) , result ) |
testMaakentityConvertorDatumGeboorte ( ) { nl . bzk . migratiebrp . test . brpnaarlo3 . adapter . ConverterContext ctx = createContext ( nl . bzk . migratiebrp . test . brpnaarlo3 . adapter . entity . IstStapelVoorkomenConverterTest . MINIMAAL ) ; converter . convertInhoudelijk ( ctx , IstStapelVoorkomenConverter . HEADER_DATUM_GEBOORTE , datumGeboorte . toString ( ) ) ; converter . maakEntity ( ctx ) ; nl . bzk . algemeenbrp . dal . domein . brp . entity . StapelVoorkomen stapelVoorkomen = ctx . getStapel ( java . lang . Integer . parseInt ( volgnummer ) ) . getStapelvoorkomens ( ) . iterator ( ) . next ( ) ; "<AssertPlaceHolder>" ; } getDatumGeboorte ( ) { return datumGeboorte ; } | org . junit . Assert . assertEquals ( datumGeboorte , stapelVoorkomen . getDatumGeboorte ( ) ) |
inlinedTestProjectsAddTheirSourceFoldersToTheMainProject ( ) { org . eclipse . xtext . xtext . wizard . TestProjectDescriptor _testProject = this . config . getRuntimeProject ( ) . getTestProject ( ) ; _testProject . setEnabled ( true ) ; this . config . setSourceLayout ( SourceLayout . MAVEN ) ; final java . util . function . Consumer < org . eclipse . xtext . xtext . wizard . SourceFolderDescriptor > _function = ( org . eclipse . xtext . xtext . wizard . SourceFolderDescriptor folder ) -> { "<AssertPlaceHolder>" ; } ; this . config . getRuntimeProject ( ) . getTestProject ( ) . getSourceFolders ( ) . forEach ( _function ) ; } getRuntimeProject ( ) { return this . runtimeProject ; } | org . junit . Assert . assertTrue ( this . config . getRuntimeProject ( ) . getSourceFolders ( ) . contains ( folder ) ) |
testClose ( ) { final java . util . concurrent . atomic . AtomicBoolean closed = new java . util . concurrent . atomic . AtomicBoolean ( false ) ; java . io . OutputStream is = new java . io . OutputStream ( ) { public void write ( int b ) throws java . io . IOException { } @ org . nucleus8583 . core . charset . spi . Override public void close ( ) throws java . io . IOException { closed . set ( true ) ; } } ; org . nucleus8583 . core . charset . spi . AsciiEncoder encoder = new org . nucleus8583 . core . charset . spi . AsciiEncoder ( is ) ; encoder . close ( ) ; "<AssertPlaceHolder>" ; } close ( ) { this . readIndex = 0 ; this . remaining = this . vlen ; } | org . junit . Assert . assertTrue ( closed . get ( ) ) |
testClose ( ) { java . sql . Connection conn = null ; try { com . ctrip . platform . dal . dao . client . DalConnection test = getConnection ( ) ; java . sql . Statement statement = test . getConn ( ) . createStatement ( ) ; java . sql . ResultSet rs = statement . executeQuery ( ( "select<sp>*<sp>from<sp>" + ( com . ctrip . platform . dal . dao . task . SqlServerTestInitializer . TABLE_NAME ) ) ) ; rs . next ( ) ; conn = test . getConn ( ) ; test . close ( ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Throwable e ) { org . junit . Assert . fail ( ) ; e . printStackTrace ( ) ; } finally { if ( conn != null ) conn . close ( ) ; } } close ( ) { try { if ( ( ( conn ) == null ) || ( conn . isClosed ( ) ) ) return ; } catch ( java . lang . Throwable e ) { logger . error ( "Restore<sp>connection<sp>isolation<sp>level<sp>failed!" , e ) ; } try { if ( ( newIsolationLevel ) != null ) conn . setTransactionIsolation ( oldIsolationLevel ) ; } catch ( java . lang . Throwable e ) { logger . error ( "Restore<sp>connection<sp>isolation<sp>level<sp>failed!" , e ) ; } try { if ( needDiscard ) { markDiscard ( conn ) ; } conn . close ( ) ; } catch ( java . lang . Throwable e ) { logger . error ( "Close<sp>connection<sp>failed!" , e ) ; } conn = null ; } | org . junit . Assert . assertTrue ( conn . isClosed ( ) ) |
classifyActivityTest ( ) { try { com . google . api . services . plus . model . Activity activity = new com . google . api . services . plus . model . Activity ( ) ; activity . setKind ( "plus#activity" ) ; java . lang . Class retClass = org . apache . streams . gplus . serializer . util . GPlusEventClassifier . detectClass ( org . apache . streams . gplus . serializer . util . GPlusEventClassifierTest . mapper . writeValueAsString ( activity ) ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception ex ) { } } detectClass ( java . lang . String ) { java . util . Objects . requireNonNull ( json ) ; com . google . common . base . Preconditions . checkArgument ( org . apache . commons . lang . StringUtils . isNotEmpty ( json ) ) ; com . fasterxml . jackson . databind . node . ObjectNode objectNode ; try { objectNode = ( ( com . fasterxml . jackson . databind . node . ObjectNode ) ( org . apache . streams . gplus . serializer . util . GPlusEventClassifier . mapper . readTree ( json ) ) ) ; } catch ( java . io . IOException ex ) { ex . printStackTrace ( ) ; return null ; } if ( ( ( objectNode . findValue ( "kind" ) ) != null ) && ( objectNode . get ( "kind" ) . toString ( ) . equals ( org . apache . streams . gplus . serializer . util . GPlusEventClassifier . ACTIVITY_IDENTIFIER ) ) ) { return com . google . api . services . plus . model . Activity . class ; } else if ( ( ( objectNode . findValue ( "kind" ) ) != null ) && ( objectNode . get ( "kind" ) . toString ( ) . equals ( org . apache . streams . gplus . serializer . util . GPlusEventClassifier . PERSON_IDENTIFIER ) ) ) { return com . google . api . services . plus . model . Person . class ; } else { return com . fasterxml . jackson . databind . node . ObjectNode . class ; } } | org . junit . Assert . assertEquals ( retClass , com . google . api . services . plus . model . Activity . class ) |
testGenerateEventDataPermutations ( ) { java . util . Map < java . lang . String , java . util . List < org . hisp . dhis . analytics . EventAnalyticsDimensionalItem > > tableRows = new java . util . LinkedHashMap ( ) ; org . hisp . dhis . common . Grid grid = new org . hisp . dhis . system . grid . ListGrid ( ) ; org . hisp . dhis . dataelement . DataElement deA = createDataElement ( 'A' ) ; deA . setValueType ( ValueType . BOOLEAN ) ; grid . addMetaData ( deA . getUid ( ) , deA ) ; org . hisp . dhis . trackedentity . TrackedEntityAttribute trackedEntityAttribute = createTrackedEntityAttribute ( 'B' ) ; org . hisp . dhis . option . OptionSet optionSet = new org . hisp . dhis . option . OptionSet ( ) ; optionSet . addOption ( new org . hisp . dhis . option . Option ( "name" , "code" ) ) ; trackedEntityAttribute . setOptionSet ( optionSet ) ; grid . addMetaData ( trackedEntityAttribute . getUid ( ) , trackedEntityAttribute ) ; java . util . List < org . hisp . dhis . analytics . EventAnalyticsDimensionalItem > objects = new java . util . ArrayList ( ) ; org . hisp . dhis . option . Option t = new org . hisp . dhis . option . Option ( ) ; t . setCode ( "1" ) ; t . setName ( "Yes" ) ; org . hisp . dhis . option . Option f = new org . hisp . dhis . option . Option ( ) ; f . setCode ( "0" ) ; f . setName ( "No" ) ; objects . add ( new org . hisp . dhis . analytics . EventAnalyticsDimensionalItem ( t , deA . getUid ( ) ) ) ; objects . add ( new org . hisp . dhis . analytics . EventAnalyticsDimensionalItem ( f , deA . getUid ( ) ) ) ; objects . add ( new org . hisp . dhis . analytics . EventAnalyticsDimensionalItem ( new org . hisp . dhis . option . Option ( "name" , "code" ) , trackedEntityAttribute . getUid ( ) ) ) ; tableRows . put ( deA . getUid ( ) , objects ) ; tableRows . put ( trackedEntityAttribute . getDimensionItem ( ) , objects ) ; java . util . List < java . util . Map < java . lang . String , org . hisp . dhis . analytics . EventAnalyticsDimensionalItem > > rowPermutations = org . hisp . dhis . analytics . event . EventAnalyticsUtils . generateEventDataPermutations ( tableRows ) ; "<AssertPlaceHolder>" ; } size ( ) { return messages . size ( ) ; } | org . junit . Assert . assertEquals ( 9 , rowPermutations . size ( ) ) |
getKeysWithoutPathToResult ( ) { "<AssertPlaceHolder>" ; } getKeys ( ) { return ( pathToResult ) == null ? null : pathToResult . split ( "/" ) ; } | org . junit . Assert . assertNull ( result . getKeys ( ) ) |
environmentVariablesHavePrecedenceOverProjectProperties ( ) { writePropertyFile ( gradleUserHomeDir , org . gradle . util . GUtil . map ( "prop" , "user<sp>value" ) ) ; writePropertyFile ( settingsDir , org . gradle . util . GUtil . map ( "prop" , "settings<sp>value" ) ) ; java . util . Map < java . lang . String , java . lang . String > projectProperties = org . gradle . util . GUtil . map ( "prop" , "project<sp>value" ) ; envProperties = org . gradle . util . GUtil . map ( ( ( IGradlePropertiesLoader . ENV_PROJECT_PROPERTIES_PREFIX ) + "prop" ) , "env<sp>value" ) ; gradlePropertiesLoader . loadProperties ( settingsDir , startParameter , systemProperties , envProperties ) ; java . util . Map < java . lang . String , java . lang . String > properties = gradlePropertiesLoader . mergeProperties ( projectProperties ) ; "<AssertPlaceHolder>" ; } get ( org . gradle . api . specs . Spec ) { return get ( elementType , spec ) ; } | org . junit . Assert . assertEquals ( "env<sp>value" , properties . get ( "prop" ) ) |
shuffleAndDealHands ( ) { bnymellon . codekatas . deckofcards . list . immutable . var jdkHands = this . jdkDeck . shuffleAndDeal ( new java . util . Random ( 1 ) , 5 , 5 ) ; bnymellon . codekatas . deckofcards . list . immutable . var acHands = this . acDeck . shuffleAndDeal ( new java . util . Random ( 1 ) , 5 , 5 ) ; "<AssertPlaceHolder>" ; } shuffleAndDeal ( java . util . Random , int , int ) { bnymellon . codekatas . deckofcards . list . immutable . var shuffled = this . shuffle ( random ) ; return this . dealHands ( shuffled , hands , cardsPerHand ) ; } | org . junit . Assert . assertEquals ( jdkHands , acHands ) |
copyConstructor_hasSameValues ( ) { array . add ( 23 ) ; com . eclipsesource . json . JsonArray copy = new com . eclipsesource . json . JsonArray ( array ) ; "<AssertPlaceHolder>" ; } values ( ) { return java . util . Collections . unmodifiableList ( values ) ; } | org . junit . Assert . assertEquals ( array . values ( ) , copy . values ( ) ) |
hasMarketplaceChanged ( ) { categorySelectionBean . lastUsedMarketplaceId = "othermid" ; "<AssertPlaceHolder>" ; } hasMarketplaceChanged ( ) { return ( ( lastUsedMarketplaceId ) == null ) || ( ! ( lastUsedMarketplaceId . equals ( ui . getMarketplaceId ( ) ) ) ) ; } | org . junit . Assert . assertTrue ( categorySelectionBean . hasMarketplaceChanged ( ) ) |
shouldCreateDatabaseDataSourceConnection ( ) { javax . sql . DataSource dataSource = mock ( javax . sql . DataSource . class ) ; java . sql . Connection connection = mock ( java . sql . Connection . class ) ; given ( dataSource . getConnection ( ) ) . willReturn ( connection ) ; this . factoryBean . setDataSource ( dataSource ) ; org . dbunit . database . DatabaseDataSourceConnection bean = this . factoryBean . getObject ( ) ; "<AssertPlaceHolder>" ; bean . getConnection ( ) . createStatement ( ) ; verify ( dataSource ) . getConnection ( ) ; } getObject ( ) { org . springframework . util . Assert . notNull ( this . dataSource , "The<sp>dataSource<sp>is<sp>required" ) ; org . dbunit . database . DatabaseDataSourceConnection dataSourceConnection = new org . dbunit . database . DatabaseDataSourceConnection ( makeTransactionAware ( this . dataSource ) , this . schema , this . username , this . password ) ; if ( ( this . databaseConfig ) != null ) { this . databaseConfig . apply ( dataSourceConnection . getConfig ( ) ) ; } return dataSourceConnection ; } | org . junit . Assert . assertNotNull ( bean ) |
shouldChangeDirection_WhenReflectedLeftDownToRightDownWhenHorizontal ( ) { com . codenjoy . dojo . pong . model . BallDirection ballDirection = new com . codenjoy . dojo . pong . model . BallDirection ( com . codenjoy . dojo . services . QDirection . LEFT_DOWN ) ; com . codenjoy . dojo . pong . model . BallDirection reflectedDirection = ballDirection . reflectedFrom ( verticalWall ) ; com . codenjoy . dojo . pong . model . BallDirection expectedDirection = new com . codenjoy . dojo . pong . model . BallDirection ( com . codenjoy . dojo . services . QDirection . RIGHT_DOWN ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == o ) return true ; if ( ( o == null ) || ( ( getClass ( ) ) != ( o . getClass ( ) ) ) ) return false ; com . codenjoy . dojo . reversi . services . Events events = ( ( com . codenjoy . dojo . reversi . services . Events ) ( o ) ) ; if ( ( count ) != ( events . count ) ) return false ; return name . equals ( events . name ) ; } | org . junit . Assert . assertTrue ( expectedDirection . equals ( reflectedDirection ) ) |
copyReaderValidWriterPosBufSz ( ) { java . lang . String probe = "A<sp>string<sp>⍅ï" ; java . io . StringWriter writer = new org . apache . maven . shared . utils . io . IOUtilTest . DontCloseStringWriter ( ) ; org . apache . maven . shared . utils . io . IOUtil . copy ( new java . io . StringReader ( probe ) , writer , 1 ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( "ReportSet{id='" + ( getId ( ) ) ) + "',<sp>reports=" ) + ( reports ) ) + "}" ; } | org . junit . Assert . assertThat ( writer . toString ( ) , org . hamcrest . CoreMatchers . is ( probe ) ) |
invalidIntegerLax ( ) { java . io . File harFile = new java . io . File ( "src/test/resources/sstoehr.invalid-integer.har" ) ; de . sstoehr . harreader . model . Har har = harReader . readFromFile ( harFile , HarReaderMode . LAX ) ; "<AssertPlaceHolder>" ; } readFromFile ( java . io . File , de . sstoehr . harreader . HarReaderMode ) { com . fasterxml . jackson . databind . ObjectMapper mapper = mapperFactory . instance ( mode ) ; try { return mapper . readValue ( har , de . sstoehr . harreader . model . Har . class ) ; } catch ( java . io . IOException e ) { throw new de . sstoehr . harreader . HarReaderException ( e ) ; } } | org . junit . Assert . assertNotNull ( har ) |
readData3 ( ) { org . mockito . Mockito . when ( filereader . fileExist ( org . mockito . Mockito . anyString ( ) ) ) . thenReturn ( true ) ; java . lang . String [ ] arr = new java . lang . String [ ] { "1.414011237411E9<sp>ACTIVE" , "1.414011237411E9<sp>ACTIVE" , "1.414011282953E9<sp>ACTIVE" , "1.414011295886E9<sp>PhonyState" , "1.414011300035E9<sp>ACTIVE" , "1.414011311924E9<sp>STANDBY" } ; org . mockito . Mockito . when ( filereader . readAllLine ( org . mockito . Mockito . anyString ( ) ) ) . thenReturn ( arr ) ; gpsEventReader . setFileReader ( filereader ) ; java . util . List < com . att . aro . core . peripheral . pojo . GpsInfo > info = gpsEventReader . readData ( "/" , 0 , 0 ) ; "<AssertPlaceHolder>" ; } size ( ) { return sessionTable . size ( ) ; } | org . junit . Assert . assertTrue ( ( ( info . size ( ) ) > 0 ) ) |
testSpecificSuppressionValue1 ( ) { net . sourceforge . pmd . Report rpt = new net . sourceforge . pmd . Report ( ) ; runTestFromString ( net . sourceforge . pmd . lang . java . SuppressWarningsTest . TEST9_VALUE1 , new net . sourceforge . pmd . FooRule ( ) , rpt , net . sourceforge . pmd . lang . LanguageRegistry . getLanguage ( JavaLanguageModule . NAME ) . getVersion ( "1.5" ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return rules . size ( ) ; } | org . junit . Assert . assertEquals ( 1 , rpt . size ( ) ) |
testExpectedHashVariableOffset ( ) { byte [ ] bytes = new byte [ 512 ] ; new java . util . Random ( 31 ) . nextBytes ( bytes ) ; for ( int i = 0 ; i <= ( bytes . length ) ; i ++ ) { int expected = org . greenrobot . essentials . hash . otherhashes . MurmurHash3Yonik . murmurhash3_x86_32 ( bytes , i , ( ( bytes . length ) - i ) , 0 ) ; checksum . reset ( ) ; checksum . update ( bytes , i , ( ( bytes . length ) - i ) ) ; int value = ( ( int ) ( checksum . getValue ( ) ) ) ; "<AssertPlaceHolder>" ; } } getValue ( ) { return ( hash ) & 4294967295L ; } | org . junit . Assert . assertEquals ( expected , value ) |
testConvertNull ( ) { "<AssertPlaceHolder>" ; } convert ( java . util . Date , org . simpleflatmapper . converter . Context ) { if ( in == null ) return null ; long time = in . getTime ( ) ; return com . google . protobuf . Timestamp . newBuilder ( ) . setSeconds ( ( time / 1000 ) ) . setNanos ( ( ( int ) ( TimeUnit . MILLISECONDS . toNanos ( ( time % 1000 ) ) ) ) ) . build ( ) ; } | org . junit . Assert . assertNull ( converter . convert ( null , null ) ) |
testRequestProvidedWorkingDirectoryShouldOverrideGlobal ( ) { logTestStart ( ) ; java . io . File tmpDir = getTempDir ( ) ; java . io . File wd = new java . io . File ( tmpDir , "workdir" ) ; java . io . File gwd = new java . io . File ( tmpDir , "global-workdir" ) ; wd . mkdirs ( ) ; gwd . mkdirs ( ) ; toDelete . add ( wd ) ; toDelete . add ( gwd ) ; org . apache . maven . shared . invoker . MavenCommandLineBuilderTest . TestCommandLineBuilder tcb = new org . apache . maven . shared . invoker . MavenCommandLineBuilderTest . TestCommandLineBuilder ( ) ; tcb . setWorkingDirectory ( gwd ) ; org . apache . maven . shared . invoker . InvocationRequest req = newRequest ( ) ; req . setBaseDirectory ( wd ) ; org . apache . maven . shared . utils . cli . Commandline cli = new org . apache . maven . shared . utils . cli . Commandline ( ) ; tcb . setEnvironmentPaths ( req , cli ) ; "<AssertPlaceHolder>" ; } getWorkingDirectory ( ) { return ( workingDir ) == null ? null : new java . io . File ( workingDir ) ; } | org . junit . Assert . assertEquals ( cli . getWorkingDirectory ( ) , wd ) |
multiDomainQueryAnd ( ) { com . b2international . index . query . Expression actual = eval ( "*<sp>{{<sp>active=false<sp>}}<sp>AND<sp>*<sp>{{<sp>term=\"clin<sp>find\"<sp>}}" ) ; com . b2international . index . query . Expression expected = com . b2international . index . query . Expressions . builder ( ) . filter ( SnomedDocument . Expressions . inactive ( ) ) . filter ( SnomedDocument . Expressions . ids ( java . util . Collections . emptySet ( ) ) ) . build ( ) ; "<AssertPlaceHolder>" ; } build ( ) { return build ( ( ( java . lang . Void ) ( null ) ) ) ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testCookie ( ) { com . woonoz . proxy . servlet . UrlRewriter rewriter = org . easymock . EasyMock . createMock ( com . woonoz . proxy . servlet . UrlRewriter . class ) ; java . lang . String headerValue = "JSESSIONID=15673243F;<sp>Path=/toto/" ; java . lang . String expectedHeader = "JSESSIONID=15673243F;<sp>Path=/test/" ; org . easymock . EasyMock . expect ( rewriter . rewriteCookie ( headerValue ) ) . andReturn ( expectedHeader ) ; org . easymock . EasyMock . replay ( rewriter ) ; com . woonoz . proxy . servlet . AbstractHeadersHandler handler = new com . woonoz . proxy . servlet . ServerHeadersHandler ( rewriter ) ; java . lang . String actualValue = handler . handleHeader ( "set-cookie" , headerValue ) ; "<AssertPlaceHolder>" ; org . easymock . EasyMock . verify ( rewriter ) ; } handleHeader ( java . lang . String , java . lang . String ) { com . woonoz . proxy . servlet . HeadersToSubstitute handler = headersToHandle . get ( toLower ( headerName ) ) ; if ( handler != null ) { return handler . handleValue ( headerValue , urlRewriter ) ; } else { return headerValue ; } } | org . junit . Assert . assertEquals ( expectedHeader , actualValue ) |
testDefect21389 ( ) { java . lang . String sql = "CREATE<sp>VIRTUAL<sp>PROCEDURE<sp>BEGIN<sp>SELECT<sp>*<sp>INTO<sp>#temptable<sp>FROM<sp>pm1.g1;<sp>INSERT<sp>INTO<sp>#temptable<sp>(e1)<sp>VALUES<sp>('a');<sp>END" ; org . teiid . query . metadata . TransformationMetadata metadata = org . teiid . query . unittest . RealMetadataFactory . example1 ( ) ; org . teiid . metadata . Column c = metadata . getElementID ( "pm1.g1.e1" ) ; c . setUpdatable ( false ) ; org . teiid . query . sql . lang . Command command = new org . teiid . query . parser . QueryParser ( ) . parseCommand ( sql ) ; org . teiid . query . resolver . QueryResolver . resolveCommand ( command , metadata ) ; org . teiid . query . validator . ValidatorReport report = org . teiid . query . validator . Validator . validate ( command , metadata ) ; "<AssertPlaceHolder>" ; } getItems ( ) { return items ; } | org . junit . Assert . assertEquals ( 0 , report . getItems ( ) . size ( ) ) |
testNs11 ( ) { org . apache . karaf . features . Repository features = unmarshalAndValidate ( "f04.xml" ) ; "<AssertPlaceHolder>" ; } unmarshalAndValidate ( java . lang . String ) { java . net . URI uri = getClass ( ) . getResource ( path ) . toURI ( ) ; return new org . apache . karaf . features . internal . service . RepositoryImpl ( uri , true ) ; } | org . junit . Assert . assertNotNull ( features ) |
testGetCapabilitiesTypeNoVersion ( ) { net . opengis . cat . csw . v_2_0_2 . GetCapabilitiesType gct = createDefaultGetCapabilitiesType ( ) ; gct . setAcceptVersions ( null ) ; net . opengis . cat . csw . v_2_0_2 . CapabilitiesType ct = null ; try { ct = org . codice . ddf . spatial . ogc . csw . catalog . endpoint . TestCswEndpoint . csw . getCapabilities ( gct ) ; } catch ( org . codice . ddf . spatial . ogc . csw . catalog . common . CswException e ) { org . junit . Assert . fail ( ( "CswException<sp>caught<sp>during<sp>getCapabilities<sp>GET<sp>request:<sp>" + ( e . getMessage ( ) ) ) ) ; } "<AssertPlaceHolder>" ; verifyOperationsMetadata ( ct ) ; verifyServiceIdentification ( ct ) ; verifyServiceProvider ( ct ) ; verifyFilterCapabilities ( ct ) ; } getCapabilities ( net . opengis . cat . csw . v_2_0_2 . GetCapabilitiesType ) { capabilitiesType = buildCapabilitiesType ( ) ; if ( ( request . getAcceptVersions ( ) ) != null ) { validateVersion ( request . getAcceptVersions ( ) . toString ( ) ) ; } java . util . List < java . lang . String > sectionList = null ; if ( ( request . getSections ( ) ) != null ) { sectionList = request . getSections ( ) . getSection ( ) ; } return buildCapabilitiesType ( sectionList ) ; } | org . junit . Assert . assertNotNull ( ct ) |
getPinpointEndpointIdTestNullEndpointIdentifier ( ) { android . content . SharedPreferences sharedPreferences = android . support . test . InstrumentationRegistry . getTargetContext ( ) . getSharedPreferences ( ( ( TEST_PP_APP_ID_1 ) + ( PP_PREFERENCES_AND_FILE_MANAGER_SUFFIX ) ) , Context . MODE_PRIVATE ) ; sharedPreferences . edit ( ) . putString ( com . amazonaws . mobileconnectors . cognitoidentityprovider . unauth . PP_UNIQUE_ID_KEY , com . amazonaws . mobileconnectors . cognitoidentityprovider . unauth . TEST_PP_ENDPOINT_2 ) . commit ( ) ; java . lang . String pp_endpoint = com . amazonaws . mobileconnectors . cognitoidentityprovider . util . CognitoPinpointSharedContext . getPinpointEndpoint ( android . support . test . InstrumentationRegistry . getTargetContext ( ) , com . amazonaws . mobileconnectors . cognitoidentityprovider . unauth . TEST_PP_APP_ID_1 , null ) ; "<AssertPlaceHolder>" ; } getPinpointEndpoint ( android . content . Context , java . lang . String , java . lang . String ) { if ( ( ( context == null ) || ( pinpointAppId == null ) ) || ( pinpointEndpointIdentifier == null ) ) { return null ; } try { final android . content . SharedPreferences pinpointPreferences = context . getSharedPreferences ( ( pinpointAppId + ( com . amazonaws . mobileconnectors . cognitoidentityprovider . util . CognitoPinpointSharedContext . PREFERENCES_AND_FILE_MANAGER_SUFFIX ) ) , Context . MODE_PRIVATE ) ; java . lang . String pinpointEndpointId = pinpointPreferences . getString ( pinpointEndpointIdentifier , null ) ; if ( pinpointEndpointId == null ) { pinpointEndpointId = java . util . UUID . randomUUID ( ) . toString ( ) ; final android . content . SharedPreferences . Editor pinpointSharedPrefsEditor = pinpointPreferences . edit ( ) ; pinpointSharedPrefsEditor . putString ( pinpointEndpointIdentifier , pinpointEndpointId ) . apply ( ) ; } return pinpointEndpointId ; } catch ( java . lang . Exception e ) { com . amazonaws . mobileconnectors . cognitoidentityprovider . util . CognitoPinpointSharedContext . LOGGER . error ( "Error<sp>while<sp>reading<sp>from<sp>SharedPreferences" , e ) ; return null ; } } | org . junit . Assert . assertNull ( pp_endpoint ) |
deveObterDadosComoFoiSetado ( ) { final com . fincatto . documentofiscal . nfe310 . classes . cadastro . NFRetornoConsultaCadastro retornoConsultaCadastro = new com . fincatto . documentofiscal . nfe310 . classes . cadastro . NFRetornoConsultaCadastro ( ) ; final com . fincatto . documentofiscal . nfe310 . classes . cadastro . NFRetornoConsultaCadastroDados dados = new com . fincatto . documentofiscal . nfe310 . classes . cadastro . NFRetornoConsultaCadastroDados ( ) ; retornoConsultaCadastro . setDados ( dados ) ; "<AssertPlaceHolder>" ; } getDados ( ) { return this . dados ; } | org . junit . Assert . assertEquals ( dados , retornoConsultaCadastro . getDados ( ) ) |
testForCorrectSeparatorBytes5 ( ) { java . lang . Object [ ] o1 = new java . lang . Object [ ] { "a" , "b" , null , null } ; java . lang . Object [ ] o2 = new java . lang . Object [ ] { null , "c" , "d" , "e" } ; org . apache . phoenix . schema . types . PDataType type = org . apache . phoenix . schema . types . PVarcharArray . INSTANCE ; org . apache . phoenix . schema . types . PDataType base = org . apache . phoenix . schema . types . PVarchar . INSTANCE ; org . apache . phoenix . schema . types . PhoenixArray arr1 = new org . apache . phoenix . schema . types . PhoenixArray ( base , o1 ) ; org . apache . phoenix . schema . types . PhoenixArray arr2 = new org . apache . phoenix . schema . types . PhoenixArray ( base , o2 ) ; org . apache . phoenix . expression . LiteralExpression array1Literal ; org . apache . phoenix . expression . LiteralExpression array2Literal ; array1Literal = org . apache . phoenix . expression . LiteralExpression . newConstant ( arr1 , type , null , null , SortOrder . DESC , Determinism . ALWAYS ) ; array2Literal = org . apache . phoenix . expression . LiteralExpression . newConstant ( arr2 , type , null , null , SortOrder . DESC , Determinism . ALWAYS ) ; java . util . List < org . apache . phoenix . expression . Expression > expressions = com . google . common . collect . Lists . newArrayList ( ( ( org . apache . phoenix . expression . Expression ) ( array1Literal ) ) ) ; expressions . add ( array2Literal ) ; org . apache . phoenix . expression . Expression arrayConcatFunction = new org . apache . phoenix . expression . function . ArrayConcatFunction ( expressions ) ; org . apache . hadoop . hbase . io . ImmutableBytesWritable ptr = new org . apache . hadoop . hbase . io . ImmutableBytesWritable ( ) ; arrayConcatFunction . evaluate ( null , ptr ) ; byte [ ] expected = new byte [ ] { - 98 , - 1 , - 99 , - 1 , 0 , - 3 , - 100 , - 1 , - 101 , - 1 , - 102 , - 1 , - 1 , - 1 , - 128 , 1 , - 128 , 3 , - 128 , 5 , - 128 , 5 , - 128 , 5 , - 128 , 7 , - 128 , 9 , - 128 , 11 , 0 , 0 , 0 , 14 , 0 , 0 , 0 , 8 , 1 } ; "<AssertPlaceHolder>" ; } get ( ) { return this . tableName ; } | org . junit . Assert . assertArrayEquals ( expected , ptr . get ( ) ) |
buildSubQueryOneCorpus ( ) { java . lang . String expected = "" + ( ( ( ( "SELECT<sp>DISTINCT<sp>c1.id<sp>" + "FROM<sp>corpus<sp>AS<sp>c1,<sp>corpus<sp>AS<sp>c2<sp>" ) + "WHERE<sp>c1.pre<sp>>=<sp>c2.pre<sp>" ) + "AND<sp>c1.post<sp><=<sp>c2.post<sp>" ) + "AND<sp>c2.id<sp>IN<sp>(<sp>23<sp>)" ) ; corpusList = java . util . Arrays . asList ( 23L ) ; "<AssertPlaceHolder>" ; } buildSubQuery ( java . util . List , java . util . List ) { java . lang . StringBuffer sb = new java . lang . StringBuffer ( ) ; sb . append ( "corpus_annotation" 4 ) ; if ( ! ( corpusList . isEmpty ( ) ) ) { sb . append ( ",<sp>" ) ; sb . append ( "corpus<sp>AS<sp>c2" ) ; } for ( int i = 1 ; i <= ( metaData . size ( ) ) ; ++ i ) { sb . append ( ",<sp>" ) ; sb . append ( "corpus_annotation<sp>AS<sp>corpus_annotation" ) ; sb . append ( i ) ; } if ( hasCorpusSelection ( corpusList , metaData ) ) sb . append ( "corpus_annotation" 5 ) ; java . util . List < java . lang . String > conditions = new java . util . ArrayList ( ) ; if ( ! ( corpusList . isEmpty ( ) ) ) { conditions . add ( "corpus_annotation" 8 ) ; conditions . add ( "corpus_annotation" 2 ) ; conditions . add ( "c2.id<sp>IN<sp>(<sp>:corpusList<sp>)" . replace ( ":corpusList" , org . apache . commons . lang3 . StringUtils . join ( corpusList , ",<sp>" ) ) ) ; } for ( int i = 1 ; i <= ( metaData . size ( ) ) ; ++ i ) { annis . model . QueryAnnotation annotation = metaData . get ( ( i - 1 ) ) ; if ( ( annotation . getNamespace ( ) ) != null ) conditions . add ( ( ( ( ( "corpus_annotation" + i ) + "corpus_annotation" 6 ) + ( annotation . getNamespace ( ) ) ) + "corpus_annotation" 0 ) ) ; conditions . add ( ( ( ( ( "corpus_annotation" + i ) + "corpus_annotation" 7 ) + ( annotation . getName ( ) ) ) + "corpus_annotation" 0 ) ) ; if ( ( annotation . getValue ( ) ) != null ) { java . lang . String value = annotation . getValue ( ) ; if ( ( ( annotation . getTextMatching ( ) ) == ( QueryNode . TextMatching . REGEXP_EQUAL ) ) || ( ( annotation . getTextMatching ( ) ) == ( QueryNode . TextMatching . REGEXP_NOT_EQUAL ) ) ) { value = ( "^" + value ) + "corpus_annotation" 1 ; } conditions . add ( ( ( ( ( ( ( "corpus_annotation" + i ) + ".value<sp>" ) + ( annotation . getTextMatching ( ) . sqlOperator ( ) ) ) + "corpus_annotation" 3 ) + value ) + "corpus_annotation" 0 ) ) ; } conditions . add ( ( ( "corpus_annotation" + i ) + ".corpus_ref<sp>=<sp>c1.id" ) ) ; } sb . append ( org . apache . commons . lang3 . StringUtils . join ( conditions , "<sp>AND<sp>" ) ) ; return sb . toString ( ) ; } | org . junit . Assert . assertEquals ( expected , strategy . buildSubQuery ( corpusList , metaData ) ) |
testTexts ( ) { com . ctrip . platform . dal . dao . sqlbuilder . AbstractFreeSqlBuilder test = create ( ) ; test . append ( texts ( com . ctrip . platform . dal . dao . sqlbuilder . AbstractFreeSqlBuilderTest . template , com . ctrip . platform . dal . dao . sqlbuilder . AbstractFreeSqlBuilderTest . template ) ) ; "<AssertPlaceHolder>" ; } build ( ) { java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; for ( com . ctrip . platform . dal . dao . sqlbuilder . Clause c : list ) sb . append ( c . build ( ) ) ; return sb . toString ( ) ; } | org . junit . Assert . assertEquals ( ( ( ( com . ctrip . platform . dal . dao . sqlbuilder . AbstractFreeSqlBuilderTest . template ) + "<sp>" ) + ( com . ctrip . platform . dal . dao . sqlbuilder . AbstractFreeSqlBuilderTest . template ) ) , test . build ( ) ) |
testAddToFacetResultListContainsCorrectValue ( ) { org . dspace . app . rest . model . SearchFacetValueRest searchFacetValueRest = new org . dspace . app . rest . model . SearchFacetValueRest ( ) ; facetResultsRest . setFacetEntry ( new org . dspace . app . rest . model . SearchFacetEntryRest ( "test" ) ) ; facetResultsRest . addToFacetResultList ( searchFacetValueRest ) ; "<AssertPlaceHolder>" ; } getFacetResultList ( ) { return facetEntry . getValues ( ) ; } | org . junit . Assert . assertEquals ( searchFacetValueRest , facetResultsRest . getFacetResultList ( ) . get ( 0 ) ) |
testHeadersEquals ( ) { org . codehaus . httpcache4j . Headers h = new org . codehaus . httpcache4j . Headers ( ) . add ( HeaderConstants . ACCEPT , "application/xhtml" ) . add ( "Foo" , "Bar" ) . add ( HeaderConstants . AGE , "23" ) ; org . codehaus . httpcache4j . Headers h2 = new org . codehaus . httpcache4j . Headers ( ) . add ( HeaderConstants . ACCEPT , "application/xhtml" ) . add ( "Foo" , "Bar" ) . add ( HeaderConstants . AGE , "23" ) ; "<AssertPlaceHolder>" ; } add ( java . lang . String , java . lang . String ) { headers = headers . add ( key , value ) ; } | org . junit . Assert . assertEquals ( h , h2 ) |
testCreateSettingsPanel ( ) { System . out . println ( "createSettingsPanel" ) ; kg . apc . jmeter . graphs . AbstractGraphPanelVisualizer instance = new kg . apc . jmeter . graphs . AbstractGraphPanelVisualizerTest . AbstractGraphPanelVisualizerImpl ( ) ; kg . apc . jmeter . vizualizers . JSettingsPanel result = instance . createSettingsPanel ( ) ; "<AssertPlaceHolder>" ; } createSettingsPanel ( ) { return new kg . apc . jmeter . vizualizers . JSettingsPanel ( this , ( ( ( ( ( ( ( JSettingsPanel . TIMELINE_OPTION ) | ( JSettingsPanel . GRADIENT_OPTION ) ) | ( JSettingsPanel . FINAL_ZEROING_OPTION ) ) | ( JSettingsPanel . LIMIT_POINT_OPTION ) ) | ( JSettingsPanel . MAXY_OPTION ) ) | ( JSettingsPanel . RELATIVE_TIME_OPTION ) ) | ( JSettingsPanel . MARKERS_OPTION ) ) ) ; } | org . junit . Assert . assertNotNull ( result ) |
testNoFailureOnNonStrict ( ) { java . io . FileOutputStream writer = new java . io . FileOutputStream ( testPath , true ) ; writeLine ( writer , "Lengthy" ) ; writeLine ( writer , "Normal" ) ; writeLine ( writer , "Short" ) ; writeLine ( writer , "Normal" ) ; writeLine ( writer , "Tiny" ) ; writer . close ( ) ; java . util . List < cascading . tuple . Tuple > expected = new java . util . ArrayList < cascading . tuple . Tuple > ( ) ; expected . addAll ( com . liveramp . cascading_ext . TestFixedWidthTextScheme . EXPECTED ) ; expected . add ( new cascading . tuple . Tuple ( "Nor" , "m" , "al" ) ) ; expected . add ( new cascading . tuple . Tuple ( "Nor" , "m" , "al" ) ) ; cascading . tap . Tap in = new cascading . tap . hadoop . Hfs ( new com . liveramp . cascading_ext . FixedWidthTextScheme ( com . liveramp . cascading_ext . TestFixedWidthTextScheme . FIELDS , com . liveramp . cascading_ext . TestFixedWidthTextScheme . COLUMN_WIDTHS , com . liveramp . cascading_ext . TestFixedWidthTextScheme . CHARSET , false ) , testPath ) ; java . util . List < cascading . tuple . Tuple > tuplesRead = getAllTuples ( in ) ; "<AssertPlaceHolder>" ; } getAllTuples ( cascading . tap . Tap ) { java . util . List < cascading . tuple . Tuple > ret = com . google . common . collect . Lists . newArrayList ( ) ; cascading . tuple . TupleEntryIterator tupleEntryIterator = sink . openForRead ( com . liveramp . cascading_ext . CascadingUtil . get ( ) . getFlowProcess ( ) ) ; while ( tupleEntryIterator . hasNext ( ) ) { ret . add ( new cascading . tuple . Tuple ( tupleEntryIterator . next ( ) . getTuple ( ) ) ) ; } return ret ; } | org . junit . Assert . assertEquals ( expected , tuplesRead ) |
testAMLaunchWithCryptoMaterial ( ) { conf . setBoolean ( CommonConfigurationKeys . IPC_SERVER_SSL_ENABLED , true ) ; conf . setBoolean ( YarnConfiguration . RM_JWT_ENABLED , false ) ; java . util . concurrent . atomic . AtomicBoolean testPass = new java . util . concurrent . atomic . AtomicBoolean ( true ) ; org . apache . hadoop . yarn . server . resourcemanager . MockRM rm = new org . apache . hadoop . yarn . server . resourcemanager . TestApplicationMasterLauncher . TestCryptoMockRM ( conf , testPass ) ; rm . start ( ) ; org . apache . hadoop . yarn . server . resourcemanager . MockNM nm = rm . registerNode ( "127.0.0.1:1337" , ( 15 * 1024 ) ) ; org . apache . hadoop . yarn . server . resourcemanager . rmapp . RMApp app = rm . submitApp ( 1024 ) ; nm . nodeHeartbeat ( true ) ; org . apache . hadoop . yarn . server . resourcemanager . rmapp . attempt . RMAppAttempt appAttempt = app . getCurrentAppAttempt ( ) ; org . apache . hadoop . yarn . server . resourcemanager . MockAM am = rm . sendAMLaunched ( appAttempt . getAppAttemptId ( ) ) ; am . registerAppAttempt ( true ) ; nm . nodeHeartbeat ( true ) ; "<AssertPlaceHolder>" ; rm . stop ( ) ; } get ( ) { return count ; } | org . junit . Assert . assertTrue ( testPass . get ( ) ) |
testDebugString ( ) { for ( byte ton : org . smpp . pdu . AddressTest . TONS ) { for ( byte npi : org . smpp . pdu . AddressTest . NPIS ) { for ( int len = 1 ; len <= ( org . smpp . pdu . AddressTest . MAX_ADDRESS_LENGTH ) ; len ++ ) { java . lang . String a = org . smpp . pdu . AddressTest . address ( len ) ; address = new org . smpp . pdu . Address ( ton , npi , a , ( len + 1 ) ) ; java . lang . String s = ( ( ( ( ( "(addr:<sp>" + ( java . lang . Integer . toString ( ton ) ) ) + "<sp>" ) + ( java . lang . Integer . toString ( npi ) ) ) + "<sp>" ) + a ) + ")<sp>" ; "<AssertPlaceHolder>" ; } } } } debugString ( ) { java . lang . String dbgs = "(data_resp:<sp>" ; dbgs += super . debugString ( ) ; dbgs += getMessageId ( ) ; dbgs += "<sp>" ; dbgs += debugStringOptional ( ) ; dbgs += ")<sp>" ; return dbgs ; } | org . junit . Assert . assertEquals ( s , address . debugString ( ) ) |
addIndexFrontRemovedTest ( ) { org . threadly . concurrent . collections . ConcurrentArrayList . DataSet < java . lang . Integer > result = org . threadly . concurrent . collections . ConcurrentArrayListDataSetTest . removedFromFront . add ( 5 , 100 ) . add ( 0 , 200 ) ; java . lang . Integer [ ] expectedResult = new java . lang . Integer [ ] { 200 , 1 , 2 , 3 , 4 , 5 , 100 , 6 , 7 , 8 , 9 } ; "<AssertPlaceHolder>" ; } makeDataSet ( java . lang . Object [ ] , int , int ) { return new org . threadly . concurrent . collections . ConcurrentArrayList . DataSet ( dataArray , startPosition , endPosition , 0 , 0 ) ; } | org . junit . Assert . assertTrue ( result . equals ( org . threadly . concurrent . collections . ConcurrentArrayListDataSetTest . makeDataSet ( expectedResult , 0 , expectedResult . length ) ) ) |
testCheckPreFlightRequestTypeNoACRM ( ) { com . erudika . para . utils . filters . MockHttpServletRequest request = new com . erudika . para . utils . filters . MockHttpServletRequest ( ) ; request . setHeader ( CORSFilter . REQUEST_HEADER_ORIGIN , TestConfigs . HTTP_TOMCAT_APACHE_ORG ) ; request . setMethod ( "OPTIONS" ) ; com . erudika . para . utils . filters . CORSFilter corsFilter = new com . erudika . para . utils . filters . CORSFilter ( ) ; corsFilter . init ( com . erudika . para . utils . filters . TestConfigs . getDefaultFilterConfig ( ) ) ; com . erudika . para . utils . filters . CORSFilter . CORSRequestType requestType = corsFilter . checkRequestType ( request ) ; "<AssertPlaceHolder>" ; } checkRequestType ( javax . servlet . http . HttpServletRequest ) { com . erudika . para . utils . filters . CORSFilter . CORSRequestType requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . INVALID_CORS ; if ( request == null ) { throw new java . lang . IllegalArgumentException ( "HttpServletRequest<sp>object<sp>is<sp>null" ) ; } java . lang . String originHeader = request . getHeader ( com . erudika . para . utils . filters . CORSFilter . REQUEST_HEADER_ORIGIN ) ; if ( originHeader != null ) { if ( originHeader . isEmpty ( ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . INVALID_CORS ; } else if ( ! ( com . erudika . para . utils . filters . CORSFilter . isValidOrigin ( originHeader ) ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . INVALID_CORS ; } else { java . lang . String method = org . apache . commons . lang3 . StringUtils . trimToEmpty ( request . getMethod ( ) ) ; if ( com . erudika . para . utils . filters . CORSFilter . HTTP_METHODS . contains ( method ) ) { if ( "OPTIONS" . equals ( method ) ) { java . lang . String accessControlRequestMethodHeader = request . getHeader ( com . erudika . para . utils . filters . CORSFilter . REQUEST_HEADER_ACCESS_CONTROL_REQUEST_METHOD ) ; if ( org . apache . commons . lang3 . StringUtils . isNotBlank ( accessControlRequestMethodHeader ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . PRE_FLIGHT ; } else if ( org . apache . commons . lang3 . StringUtils . isWhitespace ( accessControlRequestMethodHeader ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . INVALID_CORS ; } else { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . ACTUAL ; } } else if ( ( "GET" . equals ( method ) ) || ( "HEAD" . equals ( method ) ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . SIMPLE ; } else if ( "POST" . equals ( method ) ) { java . lang . String contentType = request . getContentType ( ) ; if ( contentType != null ) { contentType = contentType . toLowerCase ( ) . trim ( ) ; if ( com . erudika . para . utils . filters . CORSFilter . SIMPLE_HTTP_REQUEST_CONTENT_TYPE_VALUES . contains ( contentType ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . SIMPLE ; } else { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . ACTUAL ; } } } else if ( com . erudika . para . utils . filters . CORSFilter . COMPLEX_HTTP_METHODS . contains ( method ) ) { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . ACTUAL ; } } } } else { requestType = com . erudika . para . utils . filters . CORSFilter . CORSRequestType . NOT_CORS ; } return requestType ; } | org . junit . Assert . assertEquals ( CORSFilter . CORSRequestType . ACTUAL , requestType ) |
testCreanteStockQueueAndAddListner ( ) { vn . com . vndirect . exchangesimulator . datastorage . queue . AbstractQueueService queueService = createQueue ( "VND" ) ; vn . com . vndirect . exchangesimulator . datastorage . queue . AbstractQueueService queueService1 = stockQueueGenerator . getQueueByStock ( "VND" ) ; "<AssertPlaceHolder>" ; } getQueueByStock ( java . lang . String ) { return queues . get ( stock ) ; } | org . junit . Assert . assertEquals ( queueService , queueService1 ) |
testVerwezenObjectBestaatNiet ( ) { nl . bzk . brp . model . bericht . kern . PersoonBericht persoon1 = new nl . bzk . brp . model . bericht . kern . PersoonBericht ( ) ; persoon1 . setCommunicatieID ( "p1" ) ; persoon1 . setReferentieID ( "bestaatniet" ) ; communicatieIds . put ( persoon1 ) ; java . util . List < nl . bzk . brp . model . validatie . Melding > meldingen = brby9906 . executeer ( bericht ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , meldingen . size ( ) ) |
getHandlerForUsesNamespace ( ) { org . w3c . dom . Element el = org . apache . shindig . common . xml . XmlUtil . parseSilent ( "<Tag<sp>xmlns='#nottest'/>" ) ; "<AssertPlaceHolder>" ; } getHandlerFor ( org . apache . shindig . gadgets . templates . tags . NSName ) { org . apache . shindig . gadgets . templates . tags . TagHandler handler ; for ( org . apache . shindig . gadgets . templates . TagRegistry registry : registries ) { handler = registry . getHandlerFor ( name ) ; if ( handler != null ) { return handler ; } } return null ; } | org . junit . Assert . assertNull ( registry . getHandlerFor ( el ) ) |
testRunQuery ( ) { java . lang . String testResponse = "hello" ; org . axonframework . queryhandling . QueryMessage < java . lang . String , java . lang . String > testQueryMessage = new org . axonframework . queryhandling . GenericQueryMessage ( testResponse , org . axonframework . messaging . responsetypes . ResponseTypes . instanceOf ( java . lang . String . class ) ) ; java . lang . Object result = testSubject . handle ( testQueryMessage ) ; "<AssertPlaceHolder>" ; } handle ( java . util . List ) { events . stream ( ) . filter ( filter :: test ) . forEach ( ( event ) -> channel . send ( transform ( event ) ) ) ; } | org . junit . Assert . assertEquals ( testResponse , result ) |
shouldConvertIdToString ( ) { org . openkilda . model . PathId pathId = new org . openkilda . model . PathId ( "test_path_id" ) ; java . lang . String graphObject = new org . openkilda . persistence . converters . PathIdConverter ( ) . toGraphProperty ( pathId ) ; "<AssertPlaceHolder>" ; } getId ( ) { return id ; } | org . junit . Assert . assertEquals ( pathId . getId ( ) , graphObject ) |
testValidOptions ( ) { java . lang . String [ ] args = new java . lang . String [ ] { "--no-cache" , "//some/target" } ; java . lang . String [ ] expectedArgs = new java . lang . String [ ] { "//some/target" } ; "<AssertPlaceHolder>" ; } testWithArgs ( java . lang . String [ ] ) { com . facebook . buck . cli . RunCommand command = new com . facebook . buck . cli . RunCommand ( ) ; com . facebook . buck . cli . AdditionalOptionsCmdLineParser parser = com . facebook . buck . cli . CmdLineParserFactory . create ( command ) ; parser . parseArgument ( args ) ; return command . getArguments ( ) . toArray ( new java . lang . String [ 0 ] ) ; } | org . junit . Assert . assertArrayEquals ( expectedArgs , testWithArgs ( args ) ) |
testTableInfoXML ( ) { org . apache . hadoop . hbase . rest . client . Response response = org . apache . hadoop . hbase . rest . TestTableResource . client . get ( ( ( "/" + ( org . apache . hadoop . hbase . rest . TestTableResource . TABLE ) ) + "/regions" ) , Constants . MIMETYPE_XML ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hbase . rest . model . TableInfoModel model = ( ( org . apache . hadoop . hbase . rest . model . TableInfoModel ) ( org . apache . hadoop . hbase . rest . TestTableResource . context . createUnmarshaller ( ) . unmarshal ( new java . io . ByteArrayInputStream ( response . getBody ( ) ) ) ) ) ; checkTableInfo ( model ) ; } getCode ( ) { return this . code ; } | org . junit . Assert . assertEquals ( response . getCode ( ) , 200 ) |
testApplyFalse2 ( ) { com . liferay . dynamic . data . mapping . form . evaluator . internal . function . BetweenFunction betweenFunction = new com . liferay . dynamic . data . mapping . form . evaluator . internal . function . BetweenFunction ( ) ; boolean result = betweenFunction . apply ( BigDecimal . ONE , new java . math . BigDecimal ( 2 ) , BigDecimal . TEN ) ; "<AssertPlaceHolder>" ; } apply ( java . lang . String , java . lang . String , java . lang . String ) { if ( ( _ddmExpressionFieldAccessor ) == null ) { return false ; } try { com . liferay . dynamic . data . mapping . data . provider . DDMDataProviderRequest . Builder builder = DDMDataProviderRequest . Builder . newBuilder ( ) ; builder = builder . withDDMDataProviderId ( ddmDataProviderInstanceUUID ) ; java . util . Map < java . lang . String , java . lang . String > parameterMap = extractParameters ( paramsExpression ) ; for ( Map . Entry < java . lang . String , java . lang . String > entry : parameterMap . entrySet ( ) ) { builder = builder . withParameter ( entry . getKey ( ) , entry . getValue ( ) ) ; } com . liferay . dynamic . data . mapping . data . provider . DDMDataProviderRequest ddmDataProviderRequest = builder . build ( ) ; com . liferay . dynamic . data . mapping . data . provider . DDMDataProviderResponse ddmDataProviderResponse = ddmDataProviderInvoker . invoke ( ddmDataProviderRequest ) ; java . util . Map < java . lang . String , java . lang . String > resultMap = extractResults ( resultMapExpression ) ; setDDMFormFieldValues ( ddmDataProviderResponse , resultMap ) ; } catch ( java . lang . Exception e ) { if ( com . liferay . dynamic . data . mapping . form . evaluator . internal . function . CallFunction . _log . isDebugEnabled ( ) ) { com . liferay . dynamic . data . mapping . form . evaluator . internal . function . CallFunction . _log . debug ( e , e ) ; } } return true ; } | org . junit . Assert . assertFalse ( result ) |
whenCallingCapacity_itShouldReturnEmptyCapacity ( ) { com . graphhopper . jsprit . core . problem . solution . route . activity . End end = com . graphhopper . jsprit . core . problem . solution . route . activity . End . newInstance ( "loc" , 0.0 , 0.0 ) ; "<AssertPlaceHolder>" ; } getSize ( ) { return size ; } | org . junit . Assert . assertEquals ( 0 , end . getSize ( ) . get ( 0 ) ) |
createPluginJSONFile ( ) { java . lang . String userHome = java . lang . System . getProperty ( "user.home" ) ; java . lang . String pluginFile = ( ( ( userHome + ( java . io . File . separator ) ) + ".dockstore" ) + ( java . io . File . separator ) ) + ( io . dockstore . common . FileProvisionUtil . PLUGINS_JSON_FILENAME ) ; io . dockstore . common . FileProvisionUtil . createPluginJSONFile ( pluginFile ) ; java . io . File f = new java . io . File ( pluginFile ) ; "<AssertPlaceHolder>" ; f . delete ( ) ; } isDirectory ( ) { return directory ; } | org . junit . Assert . assertTrue ( ( ( f . exists ( ) ) && ( ! ( f . isDirectory ( ) ) ) ) ) |
testSetName ( ) { n1 . setName ( "xxx" ) ; "<AssertPlaceHolder>" ; } getName ( ) { return ( com . net2plan . gui . plugins . GUITrafficDesign . TITLE ) + "<sp>(GUI)" ; } | org . junit . Assert . assertEquals ( n1 . getName ( ) , "xxx" ) |
test2ReplaceMiddleIndex ( ) { org . antlr . v4 . tool . LexerGrammar g = new org . antlr . v4 . tool . LexerGrammar ( ( "lexer<sp>grammar<sp>T;\n" + ( ( "A<sp>:<sp>\'a\';\n" + "B<sp>:<sp>\'b\';\n" ) + "C<sp>:<sp>\'c\';\n" ) ) ) ; java . lang . String input = "abc" ; org . antlr . v4 . runtime . LexerInterpreter lexEngine = g . createLexerInterpreter ( new org . antlr . v4 . runtime . ANTLRInputStream ( input ) ) ; org . antlr . v4 . runtime . CommonTokenStream stream = new org . antlr . v4 . runtime . CommonTokenStream ( lexEngine ) ; stream . fill ( ) ; org . antlr . v4 . runtime . TokenStreamRewriter tokens = new org . antlr . v4 . runtime . TokenStreamRewriter ( stream ) ; tokens . replace ( 1 , "x" ) ; tokens . replace ( 1 , "y" ) ; java . lang . String result = tokens . getText ( ) ; java . lang . String expecting = "ayc" ; "<AssertPlaceHolder>" ; } getText ( ) { return delegate . getText ( ) ; } | org . junit . Assert . assertEquals ( expecting , result ) |
testClearJobXmls ( ) { final org . apache . oozie . fluentjob . api . action . ActionAttributesBuilder builder = getBuilderInstance ( ) ; for ( final java . lang . String jobXml : org . apache . oozie . fluentjob . api . action . TestActionAttributesBuilder . JOB_XMLS ) { builder . withJobXml ( jobXml ) ; } builder . clearJobXmls ( ) ; final org . apache . oozie . fluentjob . api . action . ActionAttributes attributes = builder . build ( ) ; final java . util . List < java . lang . String > jobXmlsList = attributes . getJobXmls ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return org . apache . oozie . event . MemoryEventQueue . currentSize . intValue ( ) ; } | org . junit . Assert . assertEquals ( 0 , jobXmlsList . size ( ) ) |
testSerialize ( ) { org . deeplearning4j . models . word2vec . VocabWord word = new org . deeplearning4j . models . word2vec . VocabWord ( 1 , "word" ) ; org . deeplearning4j . models . sequencevectors . serialization . AbstractElementFactory < org . deeplearning4j . models . word2vec . VocabWord > factory = new org . deeplearning4j . models . sequencevectors . serialization . AbstractElementFactory ( org . deeplearning4j . models . word2vec . VocabWord . class ) ; System . out . println ( ( "VocabWord<sp>JSON:<sp>" + ( factory . serialize ( word ) ) ) ) ; org . deeplearning4j . models . word2vec . VocabWord word2 = factory . deserialize ( factory . serialize ( word ) ) ; "<AssertPlaceHolder>" ; } serialize ( T extends org . deeplearning4j . models . sequencevectors . sequence . SequenceElement ) { java . lang . String json = null ; try { json = element . toJSON ( ) ; } catch ( java . lang . Exception e ) { org . deeplearning4j . models . sequencevectors . serialization . AbstractElementFactory . log . error ( "Direct<sp>serialization<sp>failed,<sp>falling<sp>back<sp>to<sp>jackson" ) ; } if ( ( json == null ) || ( json . isEmpty ( ) ) ) { org . nd4j . shade . jackson . databind . ObjectMapper mapper = org . deeplearning4j . models . sequencevectors . sequence . SequenceElement . mapper ( ) ; try { json = mapper . writeValueAsString ( element ) ; } catch ( org . deeplearning4j . models . sequencevectors . serialization . org . nd4j e ) { throw new java . lang . RuntimeException ( e ) ; } } return json ; } | org . junit . Assert . assertEquals ( word , word2 ) |
testNullMessage ( ) { org . axonframework . micrometer . EventProcessorLatencyMonitor testSubject = org . axonframework . micrometer . EventProcessorLatencyMonitor . buildMonitor ( org . axonframework . micrometer . EventProcessorLatencyMonitorTest . METER_NAME_PREFIX , meterRegistry ) ; org . axonframework . monitoring . MessageMonitor . MonitorCallback monitorCallback = testSubject . onMessageIngested ( null ) ; monitorCallback . reportSuccess ( ) ; io . micrometer . core . instrument . Gauge latencyGauge = java . util . Objects . requireNonNull ( meterRegistry . find ( ( ( org . axonframework . micrometer . EventProcessorLatencyMonitorTest . METER_NAME_PREFIX ) + ".latency" ) ) . gauge ( ) ) ; "<AssertPlaceHolder>" ; } find ( org . axonframework . modelling . saga . AssociationValue ) { java . util . Set < java . lang . String > sagasFound = new java . util . TreeSet ( ) ; sagasFound . addAll ( managedSagas . values ( ) . stream ( ) . filter ( ( saga ) -> saga . getAssociationValues ( ) . contains ( associationValue ) ) . map ( Saga :: getSagaIdentifier ) . collect ( java . util . stream . Collectors . toList ( ) ) ) ; sagasFound . addAll ( sagaStore . findSagas ( sagaType , associationValue ) ) ; return sagasFound ; } | org . junit . Assert . assertEquals ( 0 , latencyGauge . value ( ) , 0 ) |
shouldRetrieveActiveBranchesByBranchSearchId ( ) { java . lang . String officeSearchId = org . mifos . framework . util . helpers . IntegrationTestObjectMother . sampleBranchOffice ( ) . getSearchId ( ) ; java . util . List < org . mifos . reports . ui . SelectionItem > activeBranches = this . selectionItemDao . getActiveBranchesUnderUser ( officeSearchId ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( org . apache . commons . lang . StringUtils . isBlank ( loanAmount ) ) && ( org . apache . commons . lang . StringUtils . isBlank ( businessActivity ) ) ; } | org . junit . Assert . assertThat ( activeBranches . isEmpty ( ) , org . hamcrest . CoreMatchers . is ( false ) ) |
happyCase ( ) { final org . joda . time . LocalDate date = clockService . now ( ) ; org . estatio . module . agreement . dom . AgreementRole role = agreementRoleRepository . findByAgreementAndPartyAndTypeAndContainsDate ( leaseOxfTopModel , partyTopModel , artTenant , date ) ; "<AssertPlaceHolder>" ; } findByAgreementAndPartyAndTypeAndContainsDate ( org . estatio . module . agreement . dom . Agreement , org . estatio . module . party . dom . Party , org . estatio . module . agreement . dom . role . AgreementRoleType , org . joda . time . LocalDate ) { final org . joda . time . LocalDate queryDate = ( date == null ) ? new org . joda . time . LocalDate ( 1980 , 1 , 1 ) : date ; return firstMatch ( "findByAgreementAndPartyAndTypeAndContainsDate" , "agreement" , agreement , "party" , party , "type" , type , "startDate" , queryDate , "endDate" , org . incode . module . base . dom . valuetypes . LocalDateInterval . endDateFromStartDate ( queryDate ) ) ; } | org . junit . Assert . assertNotNull ( role ) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.