input stringlengths 28 18.7k | output stringlengths 39 1.69k |
|---|---|
shouldJoinValueRightEmpty ( ) { final io . confluent . ksql . structured . SchemaKStream . KsqlValueJoiner joiner = new io . confluent . ksql . structured . SchemaKStream . KsqlValueJoiner ( leftSchema , rightSchema ) ; final io . confluent . ksql . GenericRow joined = joiner . apply ( leftRow , null ) ; final java . util . List < java . lang . Object > expected = java . util . Arrays . asList ( 12L , "foobar" , null , null ) ; "<AssertPlaceHolder>" ; } getColumns ( ) { return columns ; } | org . junit . Assert . assertEquals ( expected , joined . getColumns ( ) ) |
testExtractTokenHeader ( ) { org . springframework . mock . web . MockHttpServletRequest request = new org . springframework . mock . web . MockHttpServletRequest ( HttpMethod . GET . name ( ) , "http://www.test.com" ) ; request . addHeader ( "Authorization" , ( ( ( org . springframework . security . oauth2 . common . OAuth2AccessToken . BEARER_TYPE ) + "<sp>" ) + ( OAuth2AuthenticationServiceTest . ACCESS_TOKEN_VALUE ) ) ) ; org . springframework . security . core . Authentication authentication = cookieTokenExtractor . extract ( request ) ; "<AssertPlaceHolder>" ; } getPrincipal ( ) { return principal ; } | org . junit . Assert . assertEquals ( OAuth2AuthenticationServiceTest . ACCESS_TOKEN_VALUE , authentication . getPrincipal ( ) . toString ( ) ) |
testBuild ( ) { org . spf4j . recyclable . RecyclingSupplier < org . spf4j . recyclable . impl . ExpensiveTestObject > pool = new org . spf4j . recyclable . impl . RecyclingSupplierBuilder ( 10 , new org . spf4j . recyclable . impl . ExpensiveTestObjectFactory ( ) ) . build ( ) ; org . spf4j . recyclable . impl . ObjectPoolBuilderTest . LOG . debug ( "pool<sp>=<sp>{}" , pool ) ; org . spf4j . recyclable . impl . ExpensiveTestObject object = pool . get ( ) ; org . spf4j . recyclable . impl . ObjectPoolBuilderTest . LOG . debug ( "pool<sp>=<sp>{}" , pool ) ; pool . recycle ( object , null ) ; org . spf4j . recyclable . impl . ObjectPoolBuilderTest . LOG . debug ( "pool<sp>=<sp>{}" , pool ) ; org . spf4j . recyclable . impl . ExpensiveTestObject object2 = pool . get ( ) ; "<AssertPlaceHolder>" ; pool . dispose ( ) ; } get ( ) { org . apache . avro . ImmutableSchema result = replace . get ( root ) ; if ( result == null ) { throw new java . lang . IllegalStateException ( ( ( ( "Replacement<sp>map<sp>does<sp>not<sp>contain<sp>" + ( root ) ) + ",<sp>map<sp>" ) + ( replace ) ) ) ; } return result ; } | org . junit . Assert . assertSame ( object2 , object ) |
assertTestIsNotAbleToAddPortNumberHigherThanMaxValue ( ) { io . selendroid . standalone . server . model . EmulatorPortFinder finder = anEmptyFinder ( ) ; finder . release ( ( ( io . selendroid . standalone . server . model . impl . DefaultPortFinderTest . MAX_PORT ) + 2 ) ) ; "<AssertPlaceHolder>" ; } next ( ) { if ( availablePorts . isEmpty ( ) ) { return null ; } java . util . Collections . sort ( availablePorts ) ; java . lang . Integer port = availablePorts . get ( 0 ) ; portsInUse . add ( port ) ; availablePorts . remove ( port ) ; return port ; } | org . junit . Assert . assertNull ( finder . next ( ) ) |
testUpgradeThingFromBenderToCurrent ( ) { com . freedomotic . persistence . DataUpgradeServiceImplTest . LOG . info ( ( "Upgrade<sp>a<sp>v5.5.0<sp>Thing<sp>data<sp>to<sp>v" + ( com . freedomotic . settings . Info . getVersion ( ) ) ) ) ; java . lang . String xml = org . apache . commons . io . IOUtils . toString ( this . getClass ( ) . getResourceAsStream ( "thing-5.5.0.xml" ) , "UTF-8" ) ; java . lang . String result = ( ( java . lang . String ) ( dataUpgradeService . upgrade ( com . freedomotic . model . object . EnvObject . class , xml , "5.5.0" ) ) ) ; com . thoughtworks . xstream . XStream xstream = com . freedomotic . persistence . FreedomXStream . getXstream ( ) ; com . freedomotic . model . object . EnvObject thing = ( ( com . freedomotic . model . object . EnvObject ) ( xstream . fromXML ( result ) ) ) ; com . freedomotic . persistence . DataUpgradeServiceImplTest . LOG . info ( "Loaded<sp>from<sp>upgraded<sp>XML:<sp>{}" , thing ) ; "<AssertPlaceHolder>" ; } getXstream ( ) { if ( ( com . freedomotic . persistence . FreedomXStream . xstream ) == null ) { com . freedomotic . persistence . FreedomXStream . xstream = new com . thoughtworks . xstream . XStream ( ) ; com . freedomotic . persistence . FreedomXStream . xstream . setMode ( XStream . NO_REFERENCES ) ; com . freedomotic . persistence . FreedomXStream . xstream . autodetectAnnotations ( true ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . model . object . EnvObject . class , "points" 5 ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "polygon" , com . freedomotic . model . geometry . FreedomPolygon . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . addImplicitCollection ( com . freedomotic . model . geometry . FreedomPolygon . class , "points" , "objects" 0 , com . freedomotic . model . geometry . FreedomPoint . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "points" 1 , com . freedomotic . model . geometry . FreedomEllipse . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "objects" 0 , com . freedomotic . model . geometry . FreedomPoint . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . useAttributeFor ( com . freedomotic . model . geometry . FreedomPoint . class , "points" 0 ) ; com . freedomotic . persistence . FreedomXStream . xstream . useAttributeFor ( com . freedomotic . model . geometry . FreedomPoint . class , "zone" 1 ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "points" 6 , com . freedomotic . model . geometry . FreedomShape . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "zone" 0 , com . freedomotic . model . object . Representation . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . model . ds . Config . class , "zone" 8 ) ; com . freedomotic . persistence . FreedomXStream . xstream . registerLocalConverter ( com . freedomotic . model . ds . Config . class , "zone" 7 , new com . freedomotic . persistence . TupleConverter ( ) ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "zone" 6 , com . freedomotic . model . object . EnvObject . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "zone" 9 , com . freedomotic . model . environment . Environment . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "zone" , com . freedomotic . model . environment . Zone . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . model . environment . Zone . class , "occupiers" ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . environment . Room . class , "gates" ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . environment . Room . class , "points" 4 ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . model . environment . Environment . class , "occupiers" ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . model . environment . Zone . class , "objects" ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "zone" 2 , com . freedomotic . reactions . Trigger . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "points" 8 , com . freedomotic . rules . Statement . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "command" , com . freedomotic . reactions . Command . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "zone" 3 , com . freedomotic . reactions . Reaction . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . alias ( "objects" 1 , com . freedomotic . core . Condition . class ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . reactions . Trigger . class , "zone" 4 ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . reactions . Trigger . class , "points" 3 ) ; com . freedomotic . persistence . FreedomXStream . xstream . omitField ( com . freedomotic . reactions | org . junit . Assert . assertNotNull ( thing ) |
testProcess4 ( ) { org . apache . uima . analysis_engine . AnalysisEngine seg = createEngine ( de . tudarmstadt . ukp . dkpro . core . tokit . PatternBasedTokenSegmenter . class , PatternBasedTokenSegmenter . PARAM_PATTERNS , de . tudarmstadt . ukp . dkpro . core . tokit . PatternBasedTokenSegmenterTest . PATTERNS ) ; java . lang . String content = "rent25to29point9" ; org . apache . uima . jcas . JCas cas = seg . newJCas ( ) ; cas . setDocumentText ( content ) ; new de . tudarmstadt . ukp . dkpro . core . api . segmentation . type . Token ( cas , 0 , 16 ) . addToIndexes ( ) ; seg . process ( cas ) ; java . util . List < java . lang . String > ref = asList ( "rent" , "25" , "to" , "29" , "point" , "9" ) ; java . util . List < java . lang . String > tokens = toText ( select ( cas , de . tudarmstadt . ukp . dkpro . core . api . segmentation . type . Token . class ) ) ; System . out . println ( tokens ) ; "<AssertPlaceHolder>" ; } select ( org . apache . uima . cas . CAS , java . lang . String ) { java . lang . String [ ] segments = aPath . split ( "/" , 2 ) ; java . lang . String typeName = segments [ 0 ] ; org . apache . uima . cas . Type t = aCas . getTypeSystem ( ) . getType ( typeName ) ; if ( t == null ) { throw new java . lang . IllegalStateException ( ( ( "Type<sp>[" + typeName ) + "]<sp>not<sp>found<sp>in<sp>type<sp>system" ) ) ; } java . lang . String path = ( ( segments . length ) > 1 ) ? segments [ 1 ] : "" ; return de . tudarmstadt . ukp . dkpro . core . api . featurepath . FeaturePathFactory . iterable ( de . tudarmstadt . ukp . dkpro . core . api . featurepath . FeaturePathFactory . FeaturePathIterator . create ( aCas , t , path ) ) ; } | org . junit . Assert . assertEquals ( ref , tokens ) |
testMockConfig ( ) { com . github . jsonzou . jmockdata . MockConfig mockConfig = new com . github . jsonzou . jmockdata . MockConfig ( ) . byteRange ( ( ( byte ) ( 0 ) ) , Byte . MAX_VALUE ) . shortRange ( ( ( short ) ( 0 ) ) , Short . MAX_VALUE ) . intRange ( 0 , Integer . MAX_VALUE ) . floatRange ( 0.0F , Float . MAX_EXPONENT ) . doubleRange ( 0.0 , Double . MAX_VALUE ) . longRange ( 0 , Long . MAX_VALUE ) . dateRange ( "2010-01-01" , "2020-12-30" ) . sizeRange ( 5 , 10 ) . stringSeed ( "a" , "b" , "c" ) . charSeed ( ( ( char ) ( 97 ) ) , ( ( char ) ( 98 ) ) ) ; com . github . jsonzou . jmockdata . BasicBean basicBean = com . github . jsonzou . jmockdata . JMockData . mock ( com . github . jsonzou . jmockdata . BasicBean . class , mockConfig ) ; "<AssertPlaceHolder>" ; try { com . github . jsonzou . jmockdata . JMockData . mock ( com . github . jsonzou . jmockdata . BasicBean . class , new com . github . jsonzou . jmockdata . MockConfig ( ) . dateRange ( "20100101" , "20301230" ) ) ; org . junit . Assert . fail ( ) ; } catch ( java . lang . Exception e ) { } } mock ( java . lang . Class , com . github . jsonzou . jmockdata . MockConfig ) { com . github . jsonzou . jmockdata . DataConfig config = mockConfig . getDataConfig ( clazz , null ) ; return new com . github . jsonzou . jmockdata . mocker . BaseMocker < T > ( clazz ) . mock ( config ) ; } | org . junit . Assert . assertNotNull ( basicBean ) |
testRecorderWorksAfterSerialization ( ) { com . picocontainer . DefaultPicoContainer recorded = new com . picocontainer . DefaultPicoContainer ( new com . picocontainer . behaviors . Caching ( ) ) ; recorded . addComponent ( "fruit" , "apple" ) ; com . picocontainer . DefaultPicoContainer replayed = new com . picocontainer . DefaultPicoContainer ( recorded , new com . picocontainer . behaviors . Storing ( ) ) ; com . picocontainer . DefaultPicoContainer serializedReplayed = ( ( com . picocontainer . DefaultPicoContainer ) ( serializeAndDeserialize ( replayed ) ) ) ; "<AssertPlaceHolder>" ; } getComponent ( com . googlecode . jtype . Generic ) { return null ; } | org . junit . Assert . assertEquals ( "apple" , serializedReplayed . getComponent ( "fruit" ) ) |
testFileToBytes ( ) { final java . lang . String FILE_TO_TEST = "src/test/resources/PCEStatefulCapabilityTlv1.bin" ; final java . io . File fileToCompareWith = new java . io . File ( FILE_TO_TEST ) ; try ( java . io . InputStream bytesIStream = new java . io . FileInputStream ( fileToCompareWith ) ) { final byte [ ] actualBytes = org . opendaylight . protocol . util . ByteArray . fileToBytes ( FILE_TO_TEST ) ; if ( ( fileToCompareWith . length ( ) ) > ( Integer . MAX_VALUE ) ) { throw new java . io . IOException ( "Too<sp>large<sp>file<sp>to<sp>load<sp>in<sp>byte<sp>array." ) ; } final byte [ ] expectedBytes = new byte [ ( ( int ) ( fileToCompareWith . length ( ) ) ) ] ; int offset = 0 ; int numRead = 0 ; while ( ( offset < ( expectedBytes . length ) ) && ( ( numRead = bytesIStream . read ( expectedBytes , offset , ( ( actualBytes . length ) - offset ) ) ) >= 0 ) ) { offset += numRead ; } "<AssertPlaceHolder>" ; } } fileToBytes ( java . lang . String ) { final java . io . File file = new java . io . File ( name ) ; int offset = 0 ; int numRead ; if ( ( file . length ( ) ) > ( Integer . MAX_VALUE ) ) { throw new java . io . IOException ( "Too<sp>large<sp>file<sp>to<sp>load<sp>in<sp>byte<sp>array." ) ; } final byte [ ] byteArray = new byte [ ( ( int ) ( file . length ( ) ) ) ] ; try ( java . io . FileInputStream fin = new java . io . FileInputStream ( file ) ) { while ( offset < ( byteArray . length ) ) { numRead = fin . read ( byteArray , offset , ( ( byteArray . length ) - offset ) ) ; if ( numRead >= 0 ) { offset += numRead ; } } fin . close ( ) ; } return byteArray ; } | org . junit . Assert . assertArrayEquals ( expectedBytes , actualBytes ) |
testReadSubsamplingBounds1025 ( ) { com . twelvemonkeys . imageio . plugins . jpeg . JPEGImageReader reader = createReader ( ) ; reader . setInput ( com . twelvemonkeys . imageio . plugins . jpeg . ImageIO . createImageInputStream ( getClassLoaderResource ( "/jpeg/read-error1025.jpg" ) ) ) ; com . twelvemonkeys . imageio . plugins . jpeg . ImageReadParam param = reader . getDefaultReadParam ( ) ; param . setSourceSubsampling ( 3 , 3 , 1 , 1 ) ; java . awt . image . BufferedImage image = reader . read ( 0 , param ) ; "<AssertPlaceHolder>" ; } read ( int , javax . imageio . ImageReadParam ) { java . util . Iterator < javax . imageio . ImageTypeSpecifier > imageTypes = getImageTypes ( imageIndex ) ; javax . imageio . ImageTypeSpecifier rawType = getRawImageType ( imageIndex ) ; if ( ( header . getColorMode ( ) ) != ( SGI . COLORMODE_NORMAL ) ) { processWarningOccurred ( java . lang . String . format ( "Unsupported<sp>color<sp>mode:<sp>%d,<sp>colors<sp>may<sp>look<sp>incorrect" , header . getColorMode ( ) ) ) ; } int width = getWidth ( imageIndex ) ; int height = getHeight ( imageIndex ) ; com . twelvemonkeys . imageio . plugins . sgi . BufferedImage destination = getDestination ( param , imageTypes , width , height ) ; com . twelvemonkeys . imageio . plugins . sgi . Rectangle srcRegion = new com . twelvemonkeys . imageio . plugins . sgi . Rectangle ( ) ; com . twelvemonkeys . imageio . plugins . sgi . Rectangle destRegion = new com . twelvemonkeys . imageio . plugins . sgi . Rectangle ( ) ; computeRegions ( param , width , height , destination , srcRegion , destRegion ) ; com . twelvemonkeys . imageio . plugins . sgi . WritableRaster destRaster = clipToRect ( destination . getRaster ( ) , destRegion , ( param != null ? param . getDestinationBands ( ) : null ) ) ; checkReadParamBandSettings ( param , rawType . getNumBands ( ) , destRaster . getNumBands ( ) ) ; com . twelvemonkeys . imageio . plugins . sgi . WritableRaster rowRaster = rawType . createBufferedImage ( width , 1 ) . getRaster ( ) ; com . twelvemonkeys . imageio . plugins . sgi . Raster clippedRow = clipRowToRect ( rowRaster , srcRegion , ( param != null ? param . getSourceBands ( ) : null ) , ( param != null ? param . getSourceXSubsampling ( ) : 1 ) ) ; int [ ] scanlineOffsets ; int [ ] scanlineLengths ; int compression = header . getCompression ( ) ; if ( compression == ( SGI . COMPRESSION_RLE ) ) { scanlineOffsets = new int [ height * ( header . getChannels ( ) ) ] ; scanlineLengths = new int [ height * ( header . getChannels ( ) ) ] ; imageInput . readFully ( scanlineOffsets , 0 , scanlineOffsets . length ) ; imageInput . readFully ( scanlineLengths , 0 , scanlineLengths . length ) ; } else { scanlineOffsets = null ; scanlineLengths = null ; } int xSub = ( param != null ) ? param . getSourceXSubsampling ( ) : 1 ; int ySub = ( param != null ) ? param . getSourceYSubsampling ( ) : 1 ; processImageStarted ( imageIndex ) ; for ( int c = 0 ; c < ( header . getChannels ( ) ) ; c ++ ) { com . twelvemonkeys . imageio . plugins . sgi . WritableRaster destChannel = destRaster . createWritableChild ( destRaster . getMinX ( ) , destRaster . getMinY ( ) , destRaster . getWidth ( ) , destRaster . getHeight ( ) , 0 , 0 , new int [ ] { c } ) ; com . twelvemonkeys . imageio . plugins . sgi . Raster srcChannel = clippedRow . createChild ( clippedRow . getMinX ( ) , 0 , clippedRow . getWidth ( ) , 1 , 0 , 0 , new int [ ] { c } ) ; for ( int y = 0 ; y < height ; y ++ ) { switch ( header . getBytesPerPixel ( ) ) { case 1 : byte [ ] rowDataByte = ( ( com . twelvemonkeys . imageio . plugins . sgi . DataBufferByte ) ( rowRaster . getDataBuffer ( ) ) ) . getData ( c ) ; readRowByte ( height , srcRegion , scanlineOffsets , scanlineLengths , compression , xSub , ySub , c , rowDataByte , destChannel , srcChannel , y ) ; break ; case 2 : short [ ] rowDataUShort = ( ( com . twelvemonkeys . imageio . plugins . sgi . DataBufferUShort ) ( rowRaster . getDataBuffer ( ) ) ) . getData ( c ) ; readRowUShort ( height , srcRegion , scanlineOffsets , scanlineLengths , compression , xSub , ySub , c , rowDataUShort , destChannel , srcChannel , y ) ; break ; default : throw new java . lang . AssertionError ( ) ; } processImageProgress ( ( ( ( ( 100.0F * y ) / height ) * c ) / ( header . getChannels ( ) ) ) ) ; if ( ( ( height - 1 ) - y ) < ( srcRegion . y ) ) { break ; } if ( abortRequested ( ) ) { break ; } } if ( abortRequested ( ) ) { processReadAborted ( ) ; break ; } } processImageComplete ( ) ; return destination ; } | org . junit . Assert . assertNotNull ( image ) |
whenEmployingVehicleWhereOnlyOneDistinctVehicleCanServeAParticularJob_algorithmShouldFoundDistinctSolution ( ) { final java . util . List < java . lang . Boolean > testFailed = new java . util . ArrayList < java . lang . Boolean > ( ) ; for ( int i = 0 ; i < 10 ; i ++ ) { jsprit . core . problem . VehicleRoutingProblem . Builder vrpBuilder = VehicleRoutingProblem . Builder . newInstance ( ) ; new jsprit . core . problem . io . VrpXMLReader ( vrpBuilder ) . read ( "src/test/resources/biggerProblem.xml" ) ; jsprit . core . problem . VehicleRoutingProblem vrp = vrpBuilder . build ( ) ; jsprit . core . algorithm . VehicleRoutingAlgorithm vra = new jsprit . core . algorithm . box . SchrimpfFactory ( ) . createAlgorithm ( vrp ) ; vra . setMaxIterations ( 10 ) ; try { @ jsprit . core . algorithm . SuppressWarnings ( "unused" ) java . util . Collection < jsprit . core . problem . solution . VehicleRoutingProblemSolution > solutions = vra . searchSolutions ( ) ; } catch ( jsprit . core . algorithm . recreate . NoSolutionFoundException e ) { testFailed . add ( true ) ; } } System . out . println ( ( "failed:<sp>" + ( testFailed . size ( ) ) ) ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertTrue ( testFailed . isEmpty ( ) ) |
testGetBaseType ( ) { org . teiid . language . Array array = org . teiid . dqp . internal . datamgr . TestArray . example ( null , "a" ) ; "<AssertPlaceHolder>" ; } getBaseType ( ) { checkValues ( ) ; return org . teiid . core . types . JDBCSQLTypeInfo . getSQLType ( org . teiid . core . types . DataTypeManager . getDataTypeName ( values . getClass ( ) . getComponentType ( ) ) ) ; } | org . junit . Assert . assertEquals ( java . lang . String . class , array . getBaseType ( ) ) |
test_rollSubtract ( ) { com . baeldung . gregorian . calendar . GregorianCalendarExample calendarDemo = new com . baeldung . gregorian . calendar . GregorianCalendarExample ( ) ; java . util . GregorianCalendar calendarActual = new java . util . GregorianCalendar ( 2018 , 6 , 28 ) ; java . util . GregorianCalendar calendarExpected = new java . util . GregorianCalendar ( 2018 , 6 , 28 ) ; calendarExpected . roll ( Calendar . MONTH , ( - 8 ) ) ; java . util . Date expectedDate = calendarExpected . getTime ( ) ; "<AssertPlaceHolder>" ; } rollAdd ( java . util . GregorianCalendar , int ) { calendar . roll ( GregorianCalendar . MONTH , amount ) ; return calendar . getTime ( ) ; } | org . junit . Assert . assertEquals ( expectedDate , calendarDemo . rollAdd ( calendarActual , ( - 8 ) ) ) |
testDebugPolymer ( ) { org . openscience . cdk . interfaces . IPolymer polymer = new org . openscience . cdk . debug . DebugPolymer ( ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertTrue ( ( polymer instanceof org . openscience . cdk . debug . DebugPolymer ) ) |
testRandomByteAccess ( ) { for ( int i = 0 ; i < ( buffer . capacity ( ) ) ; i ++ ) { byte value = ( ( byte ) ( random . nextInt ( ) ) ) ; buffer . setByte ( i , value ) ; } random . setSeed ( seed ) ; for ( int i = 0 ; i < ( buffer . capacity ( ) ) ; i ++ ) { byte value = ( ( byte ) ( random . nextInt ( ) ) ) ; "<AssertPlaceHolder>" ; } } getByte ( java . lang . String ) { addToDefaults ( key , null ) ; java . lang . String value = getRequired ( key ) ; return java . lang . Byte . valueOf ( value ) ; } | org . junit . Assert . assertEquals ( value , buffer . getByte ( i ) ) |
testBuild ( ) { org . lnu . is . domain . person . enrolment . subject . PersonEnrolmentSubject context = new org . lnu . is . domain . person . enrolment . subject . PersonEnrolmentSubject ( ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>PersonEnrolmentSubject<sp>e<sp>WHERE<sp>e.status=:status<sp>AND<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . person . enrolment . subject . PersonEnrolmentSubject > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setEntity ( T ) { this . entity = entity ; } | org . junit . Assert . assertEquals ( expectedQuery , actualQuery ) |
updateUda_DifferentDefinitions ( ) { uda . setUdaDefinition ( new org . oscm . domobjects . UdaDefinition ( ) ) ; try { ua . updateUda ( voUda , defSupplier , supplier ) ; } finally { verify ( ctx , times ( 1 ) ) . setRollbackOnly ( ) ; "<AssertPlaceHolder>" ; } } getUdaValue ( ) { return uda . getUdaValue ( ) ; } | org . junit . Assert . assertFalse ( voUda . getUdaValue ( ) . equals ( uda . getUdaValue ( ) ) ) |
testSimpleKeypointCorrespondence ( ) { java . util . List < org . openimaj . image . feature . local . keypoints . Keypoint > original = new java . util . ArrayList < org . openimaj . image . feature . local . keypoints . Keypoint > ( ) ; Jama . Matrix transform = Jama . Matrix . constructWithCopy ( new double [ ] [ ] { new double [ ] { 1 , 0 , 0 } , new double [ ] { 0 , 1 , 0 } , new double [ ] { 0 , 0 , 1 } } ) ; org . openimaj . image . feature . local . keypoints . Keypoint a = new org . openimaj . image . feature . local . keypoints . Keypoint ( ) ; org . openimaj . image . feature . local . keypoints . Keypoint b = new org . openimaj . image . feature . local . keypoints . Keypoint ( ) ; org . openimaj . image . feature . local . keypoints . Keypoint c = new org . openimaj . image . feature . local . keypoints . Keypoint ( ) ; a . x = 1 ; a . y = 1 ; a . ivec = new byte [ 2 ] ; a . ori = 0.5F ; a . scale = 1.0F ; b . x = 3 ; b . y = 4 ; b . ivec = new byte [ 2 ] ; b . ori = 0.5F ; b . scale = 2.0F ; c . x = 40 ; c . y = 30 ; c . ivec = new byte [ 2 ] ; c . ori = 0.5F ; c . scale = 3.0F ; original . add ( a ) ; original . add ( b ) ; original . add ( c ) ; java . util . List < org . openimaj . util . pair . Pair < org . openimaj . image . feature . local . keypoints . Keypoint > > pairs = new java . util . ArrayList < org . openimaj . util . pair . Pair < org . openimaj . image . feature . local . keypoints . Keypoint > > ( ) ; pairs . add ( new org . openimaj . util . pair . Pair < org . openimaj . image . feature . local . keypoints . Keypoint > ( a , a ) ) ; pairs . add ( new org . openimaj . util . pair . Pair < org . openimaj . image . feature . local . keypoints . Keypoint > ( b , b ) ) ; pairs . add ( new org . openimaj . util . pair . Pair < org . openimaj . image . feature . local . keypoints . Keypoint > ( c , b ) ) ; float result = org . openimaj . feature . local . matcher . KeypointCorrespondenceTestHelper . correspondance ( pairs , transform ) ; "<AssertPlaceHolder>" ; } correspondance ( java . util . List , Jama . Matrix ) { return org . openimaj . feature . local . matcher . KeypointCorrespondenceTestHelper . correspondance ( matches , transform , 1.0F ) ; } | org . junit . Assert . assertTrue ( ( result == 2 ) ) |
testStockQueueGeneratorReturnNullWhenQueueNotCreated ( ) { vn . com . vndirect . exchangesimulator . datastorage . queue . AbstractQueueService queueService = stockQueueGenerator . getQueueByStock ( "VND" ) ; "<AssertPlaceHolder>" ; } getQueueByStock ( java . lang . String ) { return queues . get ( stock ) ; } | org . junit . Assert . assertEquals ( null , queueService ) |
test_with_prefix_and_matching_pattern_and_single_host_and_replace_host ( ) { org . apache . sling . commons . testing . osgi . MockBundle bundle = new org . apache . sling . commons . testing . osgi . MockBundle ( ( - 1 ) ) ; org . apache . sling . commons . testing . osgi . MockComponentContext ctx = new org . apache . sling . commons . testing . osgi . MockComponentContext ( bundle ) ; ctx . setProperty ( "prefixes" , new java . lang . String [ ] { "CDATA" 2 } ) ; ctx . setProperty ( "attributes" , new java . lang . String [ ] { "img:src" } ) ; ctx . setProperty ( "host.pattern" , "CDATA" 1 ) ; ctx . setProperty ( "CDATA" 3 , "img:src;(\\/content\\/dam\\/.+?\\.(png|jpg))" ) ; ctx . setProperty ( "replaceHost" , true ) ; com . adobe . acs . commons . rewriter . impl . StaticReferenceRewriteTransformerFactory factory = new com . adobe . acs . commons . rewriter . impl . StaticReferenceRewriteTransformerFactory ( ) ; factory . activate ( ctx ) ; org . apache . sling . rewriter . Transformer transformer = factory . createTransformer ( ) ; transformer . setContentHandler ( handler ) ; org . xml . sax . helpers . AttributesImpl imageWithJustSrc = new org . xml . sax . helpers . AttributesImpl ( ) ; imageWithJustSrc . addAttribute ( null , "src" , null , "CDATA" , "https://www.host.com/content/dam/flower.jpg" ) ; transformer . startElement ( null , "img" , null , imageWithJustSrc ) ; verify ( handler , only ( ) ) . startElement ( isNull ( java . lang . String . class ) , eq ( "img" ) , isNull ( java . lang . String . class ) , attributesCaptor . capture ( ) ) ; java . util . List < org . xml . sax . Attributes > values = attributesCaptor . getAllValues ( ) ; "<AssertPlaceHolder>" ; } get ( java . lang . Object ) { return this . metaDataMap . get ( o ) ; } | org . junit . Assert . assertEquals ( "CDATA" 0 , values . get ( 0 ) . getValue ( 0 ) ) |
testGetTypeByClassDuplicate ( ) { setupType ( de . metas . document . refid . api . impl . ReferenceNoDAOTests . Gen1 . class ) ; setupType ( de . metas . document . refid . api . impl . ReferenceNoDAOTests . Gen1 . class ) ; de . metas . document . refid . api . impl . PlainReferenceNoDAO dao = new de . metas . document . refid . api . impl . PlainReferenceNoDAO ( ) ; "<AssertPlaceHolder>" ; dao . retrieveRefNoTypeByClass ( org . compiere . util . Env . getCtx ( ) , de . metas . document . refid . api . impl . ReferenceNoDAOTests . Gen1 . class ) ; } retrieveReferenceNoTypes ( ) { final java . util . List < de . metas . document . refid . model . I_C_ReferenceNo_Type > result = de . metas . util . Services . get ( org . adempiere . ad . dao . IQueryBL . class ) . createQueryBuilder ( de . metas . document . refid . model . I_C_ReferenceNo_Type . class ) . addOnlyActiveRecordsFilter ( ) . orderBy ( I_C_ReferenceNo_Type . COLUMNNAME_C_ReferenceNo_Type_ID ) . create ( ) . list ( ) ; return result ; } | org . junit . Assert . assertThat ( dao . retrieveReferenceNoTypes ( ) . size ( ) , org . hamcrest . Matchers . is ( 2 ) ) |
testCreateEmpty ( ) { java . util . Map < java . lang . String , java . lang . Object > reuqestMap = new java . util . HashMap ( ) ; org . apache . ambari . server . topology . SecurityConfiguration securityConfiguration = testSubject . createSecurityConfigurationFromRequest ( reuqestMap , false ) ; "<AssertPlaceHolder>" ; } createSecurityConfigurationFromRequest ( java . util . Map , boolean ) { org . apache . ambari . server . topology . SecurityConfiguration securityConfiguration ; org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . debug ( "Creating<sp>security<sp>configuration<sp>from<sp>properties:<sp>{}" , properties ) ; java . util . Map < ? , ? > securityProperties = ( ( java . util . Map < ? , ? > ) ( properties . get ( org . apache . ambari . server . topology . SecurityConfigurationFactory . SECURITY_PROPERTY_ID ) ) ) ; if ( securityProperties == null ) { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . debug ( "No<sp>security<sp>information<sp>properties<sp>provided,<sp>returning<sp>null" ) ; return null ; } java . lang . String securityTypeString = com . google . common . base . Strings . emptyToNull ( ( ( java . lang . String ) ( securityProperties . get ( org . apache . ambari . server . topology . SecurityConfigurationFactory . TYPE_PROPERTY_ID ) ) ) ) ; if ( securityTypeString == null ) { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . error ( "Type<sp>is<sp>missing<sp>from<sp>security<sp>block." ) ; throw new java . lang . IllegalArgumentException ( "Type<sp>missing<sp>from<sp>security<sp>block." ) ; } org . apache . ambari . server . state . SecurityType securityType = com . google . common . base . Enums . getIfPresent ( org . apache . ambari . server . state . SecurityType . class , securityTypeString ) . orNull ( ) ; if ( securityType == null ) { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . error ( "Unsupported<sp>security<sp>type<sp>specified:<sp>{}" , securityType ) ; throw new java . lang . IllegalArgumentException ( ( "Type<sp>missing<sp>from<sp>security<sp>block." 2 + securityTypeString ) ) ; } if ( securityType == ( org . apache . ambari . server . state . SecurityType . KERBEROS ) ) { java . lang . String descriptorReference = com . google . common . base . Strings . emptyToNull ( ( ( java . lang . String ) ( securityProperties . get ( org . apache . ambari . server . topology . SecurityConfigurationFactory . KERBEROS_DESCRIPTOR_REFERENCE_PROPERTY_ID ) ) ) ) ; java . lang . Object descriptorJsonMap = securityProperties . get ( org . apache . ambari . server . topology . SecurityConfigurationFactory . KERBEROS_DESCRIPTOR_PROPERTY_ID ) ; if ( ( descriptorReference != null ) && ( descriptorJsonMap != null ) ) { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . error ( "Type<sp>missing<sp>from<sp>security<sp>block." 1 ) ; throw new java . lang . IllegalArgumentException ( ( ( ( ( "Type<sp>missing<sp>from<sp>security<sp>block." 3 + ( org . apache . ambari . server . topology . SecurityConfigurationFactory . KERBEROS_DESCRIPTOR_PROPERTY_ID ) ) + "<sp>and<sp>" ) + ( org . apache . ambari . server . topology . SecurityConfigurationFactory . KERBEROS_DESCRIPTOR_REFERENCE_PROPERTY_ID ) ) + "Type<sp>missing<sp>from<sp>security<sp>block." 0 ) ) ; } if ( descriptorJsonMap != null ) { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . debug ( "Found<sp>embedded<sp>descriptor:<sp>{}" , descriptorJsonMap ) ; java . lang . String descriptorText = jsonSerializer . toJson ( descriptorJsonMap , java . util . Map . class ) ; if ( persistEmbeddedDescriptor ) { descriptorReference = persistKerberosDescriptor ( descriptorText ) ; } java . util . Map < ? , ? > descriptorMap = ( ( java . util . Map < ? , ? > ) ( descriptorJsonMap ) ) ; securityConfiguration = ( persistEmbeddedDescriptor ) ? org . apache . ambari . server . topology . SecurityConfiguration . withReference ( descriptorReference ) : org . apache . ambari . server . topology . SecurityConfiguration . withDescriptor ( descriptorMap ) ; } else if ( descriptorReference != null ) { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . debug ( "Found<sp>descriptor<sp>reference:<sp>{}" , descriptorReference ) ; securityConfiguration = loadSecurityConfigurationByReference ( descriptorReference ) ; } else { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . debug ( "There<sp>is<sp>no<sp>security<sp>descriptor<sp>found<sp>in<sp>the<sp>request" ) ; securityConfiguration = SecurityConfiguration . KERBEROS ; } } else { org . apache . ambari . server . topology . SecurityConfigurationFactory . LOGGER . debug ( "There<sp>is<sp>no<sp>security<sp>configuration<sp>found<sp>in<sp>the<sp>request" ) ; securityConfiguration = SecurityConfiguration . NONE ; } return securityConfiguration ; } | org . junit . Assert . assertTrue ( ( securityConfiguration == null ) ) |
testGetStaticLabel ( ) { System . out . println ( "getStaticLabel" ) ; kg . apc . jmeter . vizualizers . BytesThroughputOverTimeGui instance = new kg . apc . jmeter . vizualizers . BytesThroughputOverTimeGui ( ) ; java . lang . String result = instance . getStaticLabel ( ) ; "<AssertPlaceHolder>" ; } getStaticLabel ( ) { return kg . apc . jmeter . JMeterPluginsUtils . prefixLabel ( "Response<sp>Codes<sp>per<sp>Second" ) ; } | org . junit . Assert . assertTrue ( ( ( result . length ( ) ) > 0 ) ) |
checksWithShards ( ) { final java . util . concurrent . atomic . AtomicBoolean sawManyShards = new java . util . concurrent . atomic . AtomicBoolean ( ) ; org . apache . beam . sdk . transforms . Combine . CombineFn < java . lang . Integer , java . lang . Integer , java . lang . Integer > combineFn = new org . apache . beam . sdk . transforms . Combine . CombineFn < java . lang . Integer , java . lang . Integer , java . lang . Integer > ( ) { @ org . apache . beam . sdk . testing . Override public org . apache . beam . sdk . testing . Integer createAccumulator ( ) { return 0 ; } @ org . apache . beam . sdk . testing . Override public org . apache . beam . sdk . testing . Integer addInput ( java . lang . Integer accumulator , java . lang . Integer input ) { return accumulator + input ; } @ org . apache . beam . sdk . testing . Override public org . apache . beam . sdk . testing . Integer mergeAccumulators ( java . lang . Iterable < java . lang . Integer > accumulators ) { if ( ( org . apache . beam . vendor . guava . v20_0 . com . google . common . collect . Iterables . size ( accumulators ) ) > 2 ) { sawManyShards . set ( true ) ; } int result = 0 ; for ( int accum : accumulators ) { result += accum ; } return result ; } @ org . apache . beam . sdk . testing . Override public org . apache . beam . sdk . testing . Integer extractOutput ( java . lang . Integer accumulator ) { return accumulator ; } } ; org . apache . beam . sdk . testing . CombineFnTester . testCombineFn ( combineFn , java . util . Arrays . asList ( 1 , 1 , 1 , 1 , 1 , 2 , 2 , 2 , 2 , 2 , 3 , 3 , 3 , 3 , 3 ) , 30 ) ; "<AssertPlaceHolder>" ; } get ( ) { org . apache . beam . runners . core . construction . ArtifactServiceStager . StagingCallable . PutArtifactResponseObserver responseObserver = new org . apache . beam . runners . core . construction . ArtifactServiceStager . StagingCallable . PutArtifactResponseObserver ( ) ; org . apache . beam . vendor . grpc . v1p13p1 . io . grpc . stub . StreamObserver < org . apache . beam . model . jobmanagement . v1 . ArtifactApi . PutArtifactRequest > requestObserver = stub . putArtifact ( responseObserver ) ; org . apache . beam . model . jobmanagement . v1 . ArtifactApi . ArtifactMetadata metadata = org . apache . beam . model . jobmanagement . v1 . ArtifactApi . ArtifactMetadata . newBuilder ( ) . setName ( file . getStagingName ( ) ) . build ( ) ; org . apache . beam . model . jobmanagement . v1 . ArtifactApi . PutArtifactMetadata putMetadata = org . apache . beam . model . jobmanagement . v1 . ArtifactApi . PutArtifactMetadata . newBuilder ( ) . setMetadata ( metadata ) . setStagingSessionToken ( stagingSessionToken ) . build ( ) ; requestObserver . onNext ( org . apache . beam . model . jobmanagement . v1 . ArtifactApi . PutArtifactRequest . newBuilder ( ) . setMetadata ( putMetadata ) . build ( ) ) ; org . apache . beam . vendor . guava . v20_0 . com . google . common . hash . Hasher hasher = org . apache . beam . vendor . guava . v20_0 . com . google . common . hash . Hashing . sha256 ( ) . newHasher ( ) ; java . nio . channels . FileChannel channel = new java . io . FileInputStream ( file . getFile ( ) ) . getChannel ( ) ; java . nio . ByteBuffer readBuffer = java . nio . ByteBuffer . allocate ( bufferSize ) ; while ( ( ! ( responseObserver . isTerminal ( ) ) ) && ( ( channel . position ( ) ) < ( channel . size ( ) ) ) ) { readBuffer . clear ( ) ; channel . read ( readBuffer ) ; readBuffer . flip ( ) ; org . apache . beam . vendor . grpc . v1p13p1 . com . google . protobuf . ByteString chunk = org . apache . beam . vendor . grpc . v1p13p1 . com . google . protobuf . ByteString . copyFrom ( readBuffer ) ; hasher . putBytes ( chunk . toByteArray ( ) ) ; readBuffer . rewind ( ) ; org . apache . beam . model . jobmanagement . v1 . ArtifactApi . PutArtifactRequest request = org . apache . beam . model . jobmanagement . v1 . ArtifactApi . PutArtifactRequest . newBuilder ( ) . setData ( org . apache . beam . model . jobmanagement . v1 . ArtifactApi . ArtifactChunk . newBuilder ( ) . setData ( chunk ) . build ( ) ) . build ( ) ; requestObserver . onNext ( request ) ; } requestObserver . onCompleted ( ) ; responseObserver . awaitTermination ( ) ; if ( ( responseObserver . err . get ( ) ) != null ) { throw new java . lang . RuntimeException ( responseObserver . err . get ( ) ) ; } return metadata . toBuilder ( ) . setSha256 ( hasher . hash ( ) . toString ( ) ) . build ( ) ; } | org . junit . Assert . assertThat ( sawManyShards . get ( ) , org . hamcrest . Matchers . is ( true ) ) |
testUpdateExternalInterface ( ) { when ( alternateKeyHelper . validateStringParameter ( anyString ( ) , anyString ( ) , anyString ( ) ) ) . thenReturn ( org . finra . herd . service . impl . EXTERNAL_INTERFACE ) ; when ( alternateKeyHelper . validateStringParameter ( anyString ( ) , anyString ( ) ) ) . thenReturn ( org . finra . herd . service . impl . DISPLAY_NAME_FIELD ) ; when ( externalInterfaceDaoHelper . getExternalInterfaceEntity ( org . finra . herd . service . impl . EXTERNAL_INTERFACE ) ) . thenReturn ( org . finra . herd . service . impl . ExternalInterfaceServiceImplTest . EXTERNAL_INTERFACE_ENTITY ) ; when ( externalInterfaceDao . saveAndRefresh ( any ( org . finra . herd . model . jpa . ExternalInterfaceEntity . class ) ) ) . thenReturn ( org . finra . herd . service . impl . ExternalInterfaceServiceImplTest . EXTERNAL_INTERFACE_ENTITY ) ; org . finra . herd . model . api . xml . ExternalInterface externalInterface = externalInterfaceService . updateExternalInterface ( org . finra . herd . service . impl . ExternalInterfaceServiceImplTest . EXTERNAL_INTERFACE_KEY , org . finra . herd . service . impl . ExternalInterfaceServiceImplTest . EXTERNAL_INTERFACE_UPDATE_REQUEST ) ; "<AssertPlaceHolder>" ; verify ( alternateKeyHelper ) . validateStringParameter ( "An" , "external<sp>interface<sp>name" , org . finra . herd . service . impl . ExternalInterfaceServiceImplTest . EXTERNAL_INTERFACE_KEY . getExternalInterfaceName ( ) ) ; verify ( alternateKeyHelper ) . validateStringParameter ( "display<sp>name" , org . finra . herd . service . impl . DISPLAY_NAME_FIELD ) ; verify ( externalInterfaceDaoHelper ) . getExternalInterfaceEntity ( org . finra . herd . service . impl . EXTERNAL_INTERFACE ) ; verify ( externalInterfaceDao ) . saveAndRefresh ( any ( org . finra . herd . model . jpa . ExternalInterfaceEntity . class ) ) ; verifyNoMoreInteractionsHelper ( ) ; } getDisplayName ( ) { return displayName ; } | org . junit . Assert . assertEquals ( org . finra . herd . service . impl . DISPLAY_NAME_FIELD , externalInterface . getDisplayName ( ) ) |
testExternalInterceptor ( ) { org . springframework . context . annotation . AnnotationConfigApplicationContext context = new org . springframework . context . annotation . AnnotationConfigApplicationContext ( org . springframework . retry . annotation . EnableRetryTests . TestConfiguration . class ) ; org . springframework . retry . annotation . EnableRetryTests . InterceptableService service = context . getBean ( org . springframework . retry . annotation . EnableRetryTests . InterceptableService . class ) ; service . service ( ) ; "<AssertPlaceHolder>" ; context . close ( ) ; } getCount ( ) { return this . count ; } | org . junit . Assert . assertEquals ( 5 , service . getCount ( ) ) |
testStructTypeFor ( ) { java . util . List < java . lang . String > fieldNames = com . google . common . collect . Lists . newArrayList ( "long" 0 , "field2" , "field3" , "field4" , "field5" , "field6" ) ; java . util . List < java . lang . String > fieldTypes = com . google . common . collect . Lists . newArrayList ( "string" , "float" , "double" , "int" , "long" , "long" 1 ) ; org . apache . spark . sql . types . StructType structFromRowUtils = com . cloudera . labs . envelope . utils . RowUtils . structTypeFor ( fieldNames , fieldTypes ) ; org . apache . spark . sql . types . StructField field1 = org . apache . spark . sql . types . DataTypes . createStructField ( "long" 0 , DataTypes . StringType , true ) ; org . apache . spark . sql . types . StructField field2 = org . apache . spark . sql . types . DataTypes . createStructField ( "field2" , DataTypes . FloatType , true ) ; org . apache . spark . sql . types . StructField field3 = org . apache . spark . sql . types . DataTypes . createStructField ( "field3" , DataTypes . DoubleType , true ) ; org . apache . spark . sql . types . StructField field4 = org . apache . spark . sql . types . DataTypes . createStructField ( "field4" , DataTypes . IntegerType , true ) ; org . apache . spark . sql . types . StructField field5 = org . apache . spark . sql . types . DataTypes . createStructField ( "field5" , DataTypes . LongType , true ) ; org . apache . spark . sql . types . StructField field6 = org . apache . spark . sql . types . DataTypes . createStructField ( "field6" , DataTypes . BooleanType , true ) ; org . apache . spark . sql . types . StructType structFromAPI = org . apache . spark . sql . types . DataTypes . createStructType ( com . google . common . collect . Lists . newArrayList ( field1 , field2 , field3 , field4 , field5 , field6 ) ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( structFromRowUtils , structFromAPI ) |
testStartCustonPort ( ) { com . github . dockerjava . api . DockerClient client = mockedClient ( "1111" ) ; br . eng . rafaelsouza . imdb . DockerPostgresql postgresInstance = new br . eng . rafaelsouza . imdb . DockerPostgresql ( client , new br . eng . rafaelsouza . imdb . DatabaseConfig ( DatabaseConfig . DatabaseType . POSTGRES ) . withPort ( 1000 ) ) ; br . eng . rafaelsouza . imdb . DatabaseStatus status = postgresInstance . start ( ) ; "<AssertPlaceHolder>" ; } start ( ) { try { int databasePort = config . getPort ( ) . orElse ( 5432 ) ; com . github . dockerjava . api . model . Ports portBindings = new com . github . dockerjava . api . model . Ports ( ) ; portBindings . bind ( com . github . dockerjava . api . model . ExposedPort . tcp ( 5432 ) , com . github . dockerjava . api . model . Ports . Binding ( databasePort ) ) ; com . github . dockerjava . api . command . CreateContainerResponse createContainerResponse = dockerClient . createContainerCmd ( "postgres" ) . withPortBindings ( portBindings ) . exec ( ) ; containerId = createContainerResponse . getId ( ) ; dockerClient . startContainerCmd ( containerId ) . exec ( ) ; java . lang . Thread . sleep ( 4000 ) ; return new br . eng . rafaelsouza . imdb . DatabaseStatus ( true , containerId , databasePort ) ; } catch ( java . lang . InterruptedException ex ) { } return new br . eng . rafaelsouza . imdb . DatabaseStatus ( false ) ; } | org . junit . Assert . assertEquals ( new br . eng . rafaelsouza . imdb . DatabaseStatus ( true , "1111" , 1000 ) , status ) |
testLockingBeforeSave ( ) { org . nuxeo . ecm . core . api . DocumentModel root = session . getRootDocument ( ) ; org . nuxeo . ecm . core . api . DocumentModel doc = session . createDocumentModel ( "/" , "doc" , "File" ) ; doc = session . createDocument ( doc ) ; doc . setLock ( ) ; session . save ( ) ; nextTransaction ( ) ; doc = session . getChild ( root . getRef ( ) , "doc" ) ; "<AssertPlaceHolder>" ; } isLocked ( ) { return doc . isLocked ( ) ; } | org . junit . Assert . assertTrue ( doc . isLocked ( ) ) |
isZero ( ) { "<AssertPlaceHolder>" ; } isZero ( ) { org . junit . Assert . assertTrue ( time . isZero ( ) ) ; } | org . junit . Assert . assertTrue ( time . isZero ( ) ) |
testMatchesWithQuotes2 ( ) { com . orientechnologies . orient . core . sql . query . OSQLSynchQuery sql = new com . orientechnologies . orient . core . sql . query . OSQLSynchQuery ( "SELECT<sp>FROM<sp>matchesstuff<sp>WHERE<sp>(name<sp>matches<sp>\'\\\\Qadm\\\\E.*\'<sp>and<sp>not<sp>(<sp>name<sp>matches<sp>\'(.*)foo(.*)\'<sp>)<sp>)" ) ; com . orientechnologies . orient . core . sql . List < com . orientechnologies . orient . core . record . impl . ODocument > results = com . orientechnologies . orient . core . sql . OCommandExecutorSQLSelectTest . db . query ( sql ) ; "<AssertPlaceHolder>" ; } size ( ) { throw new java . lang . UnsupportedOperationException ( "Not<sp>implemented<sp>yet" ) ; } | org . junit . Assert . assertEquals ( results . size ( ) , 1 ) |
testFindSessions ( ) { System . out . println ( "findSessions" ) ; com . rhythm . pb . RequestProtos . SessionKey searchKey = com . rhythm . pb . RequestProtos . SessionKey . newBuilder ( ) . setKey ( "vr" ) . build ( ) ; java . util . List < com . rhythm . pb . RequestProtos . SessionBPB > session = client . findSessions ( searchKey ) ; "<AssertPlaceHolder>" ; for ( com . rhythm . pb . RequestProtos . SessionBPB s : session ) { System . out . println ( s ) ; } } findSessions ( com . rhythm . pb . RequestProtos . SessionKey ) { com . rhythm . louie . services . auth . SessionStat stat = SESSION_STATS . get ( sessionKey . getKey ( ) ) ; if ( stat != null ) { return java . util . Collections . singletonList ( stat . toPB ( ) ) ; } else { java . util . List < com . rhythm . pb . RequestProtos . SessionBPB > found = new java . util . ArrayList ( ) ; for ( java . lang . Object key : SESSION_STATS . asMap ( ) . keySet ( ) ) { if ( key . toString ( ) . startsWith ( sessionKey . getKey ( ) ) ) { stat = SESSION_STATS . get ( key . toString ( ) ) ; if ( stat != null ) { found . add ( stat . toPB ( ) ) ; } } } if ( ( found . isEmpty ( ) ) && ( ! ( isCentralAuth ( ) ) ) ) { return getAuthClient ( ) . findSessions ( sessionKey ) ; } return found ; } } | org . junit . Assert . assertNotNull ( session ) |
testBerichtMetOverrulMetOverrulbareFouten2 ( ) { nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht bericht = new nl . bzk . brp . business . dto . bijhouding . VerhuizingBericht ( ) ; nl . bzk . brp . business . dto . BerichtContext context = new nl . bzk . brp . business . dto . BerichtContext ( new nl . bzk . brp . business . dto . BerichtenIds ( 1L , 1L ) , 1 , new nl . bzk . brp . model . gedeeld . Partij ( ) , "ref" ) ; bericht . setOverruledMeldingen ( java . util . Arrays . asList ( new nl . bzk . brp . model . validatie . OverruleMelding ( MeldingCode . AUTH0001 . getNaam ( ) ) ) ) ; nl . bzk . brp . business . dto . BerichtResultaat resultaat = new nl . bzk . brp . business . dto . BerichtResultaat ( java . util . Arrays . asList ( new nl . bzk . brp . model . validatie . Melding ( nl . bzk . brp . model . validatie . SoortMelding . INFO , nl . bzk . brp . model . validatie . MeldingCode . BRAL0012 ) , new nl . bzk . brp . model . validatie . Melding ( nl . bzk . brp . model . validatie . SoortMelding . FOUT_OVERRULEBAAR , nl . bzk . brp . model . validatie . MeldingCode . AUTH0001 ) , new nl . bzk . brp . model . validatie . Melding ( nl . bzk . brp . model . validatie . SoortMelding . INFO , nl . bzk . brp . model . validatie . MeldingCode . BRAL2032 ) , new nl . bzk . brp . model . validatie . Melding ( nl . bzk . brp . model . validatie . SoortMelding . FOUT_OVERRULEBAAR , nl . bzk . brp . model . validatie . MeldingCode . AUTH0001 ) , new nl . bzk . brp . model . validatie . Melding ( nl . bzk . brp . model . validatie . SoortMelding . FOUT_OVERRULEBAAR , nl . bzk . brp . model . validatie . MeldingCode . AUTH0001 ) , new nl . bzk . brp . model . validatie . Melding ( nl . bzk . brp . model . validatie . SoortMelding . WAARSCHUWING , nl . bzk . brp . model . validatie . MeldingCode . ALG0001 ) ) ) ; bedrijfsregelValidatieStap . corrigeerVoorOverrulebareFouten ( bericht , context , resultaat ) ; "<AssertPlaceHolder>" ; } bevatVerwerkingStoppendeFouten ( ) { for ( nl . bzk . brp . model . validatie . Melding melding : getMeldingen ( ) ) { if ( ( ( melding . getSoort ( ) ) == ( nl . bzk . brp . model . algemeen . stamgegeven . ber . SoortMelding . DEBLOKKEERBAAR ) ) || ( ( melding . getSoort ( ) ) == ( nl . bzk . brp . model . algemeen . stamgegeven . ber . SoortMelding . FOUT ) ) ) { return true ; } } return false ; } | org . junit . Assert . assertEquals ( false , resultaat . bevatVerwerkingStoppendeFouten ( ) ) |
testEndExecutionListenerIsCalledOnlyOnce ( ) { org . camunda . bpm . model . bpmn . BpmnModelInstance modelInstance = org . camunda . bpm . model . bpmn . Bpmn . createExecutableProcess ( "conditionalProcessKey" ) . startEvent ( ) . userTask ( ) . camundaExecutionListenerClass ( ExecutionListener . EVENTNAME_END , org . camunda . bpm . engine . test . bpmn . event . conditional . SetVariableDelegate . class . getName ( ) ) . camundaExecutionListenerClass ( ExecutionListener . EVENTNAME_END , org . camunda . bpm . engine . test . bpmn . executionlistener . RecorderExecutionListener . class . getName ( ) ) . endEvent ( ) . done ( ) ; modelInstance = modify ( modelInstance ) . addSubProcessTo ( "conditionalProcessKey" ) . triggerByEvent ( ) . embeddedSubProcess ( ) . startEvent ( ) . interrupting ( true ) . conditionalEventDefinition ( ) . condition ( "${variable<sp>==<sp>1}" ) . conditionalEventDefinitionDone ( ) . endEvent ( ) . done ( ) ; testHelper . deploy ( modelInstance ) ; org . camunda . bpm . engine . runtime . ProcessInstance procInst = runtimeService . startProcessInstanceByKey ( "conditionalProcessKey" ) ; org . camunda . bpm . engine . task . TaskQuery taskQuery = taskService . createTaskQuery ( ) . processInstanceId ( procInst . getId ( ) ) ; taskService . complete ( taskQuery . singleResult ( ) . getId ( ) ) ; "<AssertPlaceHolder>" ; } getRecordedEvents ( ) { return org . camunda . bpm . engine . test . bpmn . executionlistener . RecorderExecutionListener . recordedEvents ; } | org . junit . Assert . assertEquals ( 1 , org . camunda . bpm . engine . test . bpmn . executionlistener . RecorderExecutionListener . getRecordedEvents ( ) . size ( ) ) |
testRepeatedExecutionJobWithRemove ( ) { org . drools . core . SessionConfiguration config = org . drools . core . SessionConfiguration . newInstance ( ) ; config . setClockType ( ClockType . REALTIME_CLOCK ) ; org . drools . core . time . TimerService timeService = org . drools . core . time . TimerServiceFactory . getTimerService ( config ) ; org . drools . core . time . Trigger trigger = new org . drools . core . time . impl . JDKTimerServiceTest . DelayedTrigger ( new long [ ] { 100 , 100 , 100 , 100 , 100 , 100 , 100 , 100 } ) ; org . drools . core . time . impl . JDKTimerServiceTest . HelloWorldJobContext ctx = new org . drools . core . time . impl . JDKTimerServiceTest . HelloWorldJobContext ( "hello<sp>world" , timeService ) ; ctx . setLimit ( 3 ) ; timeService . scheduleJob ( new org . drools . core . time . impl . JDKTimerServiceTest . HelloWorldJob ( ) , ctx , trigger ) ; java . lang . Thread . sleep ( 1000 ) ; timeService . shutdown ( ) ; "<AssertPlaceHolder>" ; } getList ( ) { return list ; } | org . junit . Assert . assertEquals ( 5 , ctx . getList ( ) . size ( ) ) |
testGetConvNetworkId ( ) { java . lang . String nwId = "network01" ; java . util . ArrayList < java . lang . String > nwList = new java . util . ArrayList < java . lang . String > ( ) { { add ( "network01" ) ; add ( "network02" ) ; } } ; org . o3project . odenos . core . component . ConversionTable conversionTable = org . powermock . api . mockito . PowerMockito . spy ( new org . o3project . odenos . core . component . ConversionTable ( ) ) ; org . powermock . api . mockito . PowerMockito . doReturn ( nwList ) . when ( conversionTable , "getNetwork" , nwId ) ; org . powermock . api . mockito . PowerMockito . doReturn ( conversionTable ) . when ( target , "conversionTable" ) ; "<AssertPlaceHolder>" ; } getConvNetworkId ( java . lang . String ) { org . o3project . odenos . component . federator . Federator . log . debug ( "" ) ; if ( networkId == null ) { return null ; } java . util . ArrayList < java . lang . String > convNetowrkId = conversionTable ( ) . getNetwork ( networkId ) ; if ( ( convNetowrkId . size ( ) ) == 0 ) { return null ; } return convNetowrkId . get ( 0 ) ; } | org . junit . Assert . assertThat ( target . getConvNetworkId ( nwId ) , org . hamcrest . CoreMatchers . is ( "network01" ) ) |
testGetQueryName ( ) { "<AssertPlaceHolder>" ; } getQueryName ( ) { return queryName ; } | org . junit . Assert . assertEquals ( queryName , instance . getQueryName ( ) ) |
getObjectDouble ( ) { org . mockito . Mockito . when ( resultSet . getObject ( 1 ) ) . thenReturn ( 0.1 ) ; org . mockito . Mockito . when ( resultSet . getDouble ( 1 ) ) . thenReturn ( 0.1 ) ; "<AssertPlaceHolder>" ; org . mockito . Mockito . verify ( resultSet ) . getDouble ( 1 ) ; } getObject ( java . lang . Class , java . sql . ResultSet , int ) { if ( type == ( java . lang . String . class ) ) { byte [ ] bytes = rs . getBytes ( index ) ; if ( bytes != null ) { try { return ( ( M ) ( new java . lang . String ( bytes , "UTF-8" ) ) ) ; } catch ( java . io . UnsupportedEncodingException e ) { throw new com . cloud . utils . exception . CloudRuntimeException ( "UnsupportedEncodingException<sp>exception<sp>while<sp>converting<sp>UTF-8<sp>data" ) ; } } else { return null ; } } else if ( type == ( int . class ) ) { return ( ( M ) ( ( java . lang . Integer ) ( rs . getInt ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Integer . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Integer ) ( rs . getInt ( index ) ) ) ) ; } } else if ( type == ( long . class ) ) { return ( ( M ) ( ( java . lang . Long ) ( rs . getLong ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Long . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Long ) ( rs . getLong ( index ) ) ) ) ; } } else if ( type == ( java . util . Date . class ) ) { final java . lang . Object data = rs . getDate ( index ) ; if ( data == null ) { return null ; } else { return ( ( M ) ( com . cloud . utils . DateUtil . parseDateString ( com . cloud . utils . db . GenericDaoBase . s_gmtTimeZone , rs . getString ( index ) ) ) ) ; } } else if ( type == ( short . class ) ) { return ( ( M ) ( ( java . lang . Short ) ( rs . getShort ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Short . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Short ) ( rs . getShort ( index ) ) ) ) ; } } else if ( type == ( boolean . class ) ) { return ( ( M ) ( ( java . lang . Boolean ) ( rs . getBoolean ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Boolean . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Boolean ) ( rs . getBoolean ( index ) ) ) ) ; } } else if ( type == ( float . class ) ) { return ( ( M ) ( ( java . lang . Float ) ( rs . getFloat ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Float . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Float ) ( rs . getFloat ( index ) ) ) ) ; } } else if ( type == ( double . class ) ) { return ( ( M ) ( ( java . lang . Double ) ( rs . getDouble ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Double . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Double ) ( rs . getDouble ( index ) ) ) ) ; } } else if ( type == ( byte . class ) ) { return ( ( M ) ( ( java . lang . Byte ) ( rs . getByte ( index ) ) ) ) ; } else if ( type == ( com . cloud . utils . db . Byte . class ) ) { if ( ( rs . getObject ( index ) ) == null ) { return null ; } else { return ( ( M ) ( ( java . lang . Byte ) ( rs . getByte ( index ) ) ) ) ; } } else if ( type == ( java . util . Calendar . class ) ) { final java . lang . Object | org . junit . Assert . assertEquals ( 0.1 , com . cloud . utils . db . GenericDaoBase . getObject ( com . cloud . utils . db . Double . class , resultSet , 1 ) , 0.1 ) |
testEcho ( ) { java . lang . String hi = redis1 . echo ( "hi" ) ; "<AssertPlaceHolder>" ; System . out . println ( java . lang . String . format ( "[CRAFT-ATOM-REDIS]<sp>(^_^)<sp><%s><sp>Case<sp>-><sp>test<sp>echo.<sp>" , io . craft . atom . test . CaseCounter . incr ( 1 ) ) ) ; } echo ( java . lang . String ) { return master ( ) . echo ( message ) ; } | org . junit . Assert . assertEquals ( "hi" , hi ) |
getAddresses ( ) { java . util . Set < java . net . InetSocketAddress > actual = trackerPool . getAddresses ( ) ; java . util . Set < java . net . InetSocketAddress > expected = new java . util . HashSet < java . net . InetSocketAddress > ( java . util . Arrays . asList ( address1 , address2 ) ) ; "<AssertPlaceHolder>" ; } getAddresses ( ) { return poolingTrackerFactory . getAddresses ( ) ; } | org . junit . Assert . assertThat ( actual , org . hamcrest . CoreMatchers . is ( expected ) ) |
testHandlesRequest_3 ( ) { final org . apache . shindig . gadgets . oauth2 . handler . CodeAuthorizationResponseHandler fixture = org . apache . shindig . gadgets . oauth2 . handler . CodeAuthorizationResponseHandlerTest . carh ; final org . apache . shindig . gadgets . oauth2 . OAuth2Accessor accessor = null ; final javax . servlet . http . HttpServletRequest request = new org . apache . shindig . gadgets . oauth2 . handler . CodeAuthorizationResponseHandlerTest . DummyHttpServletRequest ( ) ; final boolean result = fixture . handlesRequest ( accessor , request ) ; "<AssertPlaceHolder>" ; } handlesRequest ( org . apache . shindig . gadgets . oauth2 . OAuth2Accessor , javax . servlet . http . HttpServletRequest ) { if ( accessor == null ) { return false ; } else if ( request == null ) { return false ; } else if ( ( ( ! ( accessor . isValid ( ) ) ) || ( accessor . isErrorResponse ( ) ) ) || ( ! ( accessor . isRedirecting ( ) ) ) ) { return false ; } else if ( ! ( accessor . getGrantType ( ) . equalsIgnoreCase ( OAuth2Message . AUTHORIZATION ) ) ) { return false ; } return true ; } | org . junit . Assert . assertEquals ( false , result ) |
testProcessSingleNotInitialFragment ( ) { org . hawkular . apm . api . model . trace . Trace trace = new org . hawkular . apm . api . model . trace . Trace ( ) . setTraceId ( "traceId" ) . setFragmentId ( "anotherId" ) ; org . hawkular . apm . api . model . trace . Consumer c = new org . hawkular . apm . api . model . trace . Consumer ( ) ; trace . getNodes ( ) . add ( c ) ; org . hawkular . apm . server . processor . tracecompletiontime . TraceCompletionInformationInitiator initiator = new org . hawkular . apm . server . processor . tracecompletiontime . TraceCompletionInformationInitiator ( ) ; "<AssertPlaceHolder>" ; } processOneToOne ( java . lang . String , org . hawkular . apm . server . processor . zipkin . CompletionTimeProcessing ) { if ( ( completionTimeProcessing . getCompletionTime ( ) ) == null ) { return null ; } org . hawkular . apm . server . processor . zipkin . CompletionTimeDeriver . log . debugf ( "SpanTraceCompletionTimeDeriver<sp>completionTime[%s]" , completionTimeProcessing . getCompletionTime ( ) ) ; return completionTimeProcessing . getCompletionTime ( ) ; } | org . junit . Assert . assertNull ( initiator . processOneToOne ( null , trace ) ) |
test ( ) { java . lang . String qsort = com . google . common . io . Resources . toString ( com . google . common . io . Resources . getResource ( org . spf4j . zel . vm . MinMaxTest . class , "minmax.zel" ) , StandardCharsets . US_ASCII ) ; org . spf4j . zel . vm . Program p = org . spf4j . zel . vm . Program . compile ( qsort ) ; java . lang . Integer max = ( ( java . lang . Integer ) ( p . execute ( ) ) ) ; "<AssertPlaceHolder>" ; } execute ( ) { java . io . File f = outputDirectory ; if ( ( ! ( f . exists ( ) ) ) && ( ! ( f . mkdirs ( ) ) ) ) { throw new org . apache . maven . plugin . MojoExecutionException ( ( "Unable<sp>to<sp>create<sp>directory<sp>" + ( outputDirectory ) ) ) ; } java . io . File outFile = new java . io . File ( f , fileName ) ; getLog ( ) . info ( ( "Creating<sp>avdl<sp>file<sp>at<sp>" + outFile ) ) ; try ( java . io . Writer w = new java . io . OutputStreamWriter ( new java . io . FileOutputStream ( outFile ) , encoding ) ) { if ( ( ( namespace ) != null ) && ( ! ( namespace . isEmpty ( ) ) ) ) { w . write ( "@namespace(\"" ) ; w . write ( namespace ) ; w . write ( "\")\n" ) ; } w . write ( "protocol<sp>" ) ; w . write ( rootRecordName ) ; w . write ( "Protocol" ) ; w . write ( "<sp>{\n" ) ; java . util . Map < java . lang . String , java . lang . Object > record = new java . util . HashMap ( ) ; processClasses ( classes , record ) ; writeRecord ( w , namespace , rootRecordName , record ) ; w . write ( "}\n" ) ; } catch ( java . io . IOException ex ) { throw new org . apache . maven . plugin . MojoExecutionException ( "Cannot<sp>generate<sp>config<sp>description" , ex ) ; } } | org . junit . Assert . assertEquals ( 8 , ( ( int ) ( max ) ) ) |
testManageWebAppSourceControl ( ) { "<AssertPlaceHolder>" ; } runSample ( com . microsoft . azure . management . Azure ) { final java . lang . String suffix = "/coffeeshop..." 1 ; final java . lang . String app1Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/helloworld.war" 0 , 20 ) ; final java . lang . String app2Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/coffeeshop..." 4 , 20 ) ; final java . lang . String app3Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/coffeeshop" 9 , 20 ) ; final java . lang . String app4Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/helloworld.war" 8 , 20 ) ; final java . lang . String app5Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/helloworld.war" 1 , 20 ) ; final java . lang . String app6Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/helloworld.war" 1 , 20 ) ; final java . lang . String app7Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/azure-samples-appservice-helloworld/" 3 , 20 ) ; final java . lang . String app1Url = app1Name + suffix ; final java . lang . String app2Url = app2Name + suffix ; final java . lang . String app3Url = app3Name + suffix ; final java . lang . String app4Url = app4Name + suffix ; final java . lang . String app5Url = app5Name + suffix ; final java . lang . String app6Url = app6Name + suffix ; final java . lang . String app7Url = app7Name + suffix ; final java . lang . String rgName = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/coffeeshop" 8 , 24 ) ; final java . lang . String rg7Name = com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . randomResourceName ( "/azure-samples-appservice-helloworld/" 5 , 24 ) ; try { System . out . println ( ( ( ( ( "/helloworld.war" 4 + app1Name ) + "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" 0 ) + rgName ) + "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" 6 ) ) ; com . microsoft . azure . management . appservice . WebApp app1 = azure . webApps ( ) . define ( app1Name ) . withRegion ( Region . US_WEST ) . withNewResourceGroup ( rgName ) . withNewWindowsPlan ( PricingTier . STANDARD_S1 ) . withJavaVersion ( JavaVersion . JAVA_8_NEWEST ) . withWebContainer ( WebContainer . TOMCAT_8_0_NEWEST ) . create ( ) ; System . out . println ( ( "/coffeeshop" 1 + ( app1 . name ( ) ) ) ) ; com . microsoft . azure . management . samples . Utils . print ( app1 ) ; System . out . println ( ( ( "/coffeeshop..." 3 + app1Name ) + "/azure-samples-appservice-helloworld/" 6 ) ) ; com . microsoft . azure . management . samples . Utils . uploadFileToWebApp ( app1 . getPublishingProfile ( ) , "/coffeeshop" 5 , com . microsoft . azure . management . appservice . samples . ManageWebAppSourceControl . class . getResourceAsStream ( "/helloworld.war" ) ) ; System . out . println ( ( ( "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" + ( app1 . name ( ) ) ) + "/coffeeshop..." 9 ) ) ; com . microsoft . azure . management . samples . Utils . print ( app1 ) ; System . out . println ( ( ( "/helloworld.war" 2 + app1Url ) + "/coffeeshop..." 5 ) ) ; com . microsoft . azure . management . appservice . samples . ManageWebAppSourceControl . curl ( ( ( "/azure-samples-appservice-helloworld/" 7 + app1Url ) + "/helloworld.war" 6 ) ) ; com . microsoft . azure . management . resources . fluentcore . utils . SdkContext . sleep ( 5000 ) ; System . out . println ( ( ( "/coffeeshop" 7 + app1Url ) + "/coffeeshop..." 5 ) ) ; System . out . println ( com . microsoft . azure . management . appservice . samples . ManageWebAppSourceControl . curl ( ( ( "/azure-samples-appservice-helloworld/" 7 + app1Url ) + "/helloworld.war" 6 ) ) ) ; System . out . println ( ( ( ( ( "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" 1 + app2Name ) + "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" 0 ) + rgName ) + "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" 6 ) ) ; com . microsoft . azure . management . appservice . AppServicePlan plan = azure . appServices ( ) . appServicePlans ( ) . getById ( app1 . appServicePlanId ( ) ) ; com . microsoft . azure . management . appservice . WebApp app2 = azure . webApps ( ) . define ( app2Name ) . withExistingWindowsPlan ( plan ) . withExistingResourceGroup ( rgName ) . withLocalGitSourceControl ( ) . withJavaVersion ( JavaVersion . JAVA_8_NEWEST ) . withWebContainer ( WebContainer . TOMCAT_8_0_NEWEST ) . create ( ) ; System . out . println ( ( "/coffeeshop" 1 + ( app2 . name ( ) ) ) ) ; com . microsoft . azure . management . samples . Utils . print ( app2 ) ; System . out . println ( ( ( "Deployment<sp>helloworld.war<sp>to<sp>web<sp>app<sp>" 2 + app2Name ) + "<sp>through<sp>Git..." ) ) ; com . microsoft . azure . management . appservice . PublishingProfile profile = app2 . getPublishingProfile ( ) ; org . eclipse . jgit . api . Git git = org . eclipse . jgit . api . Git . init ( ) . setDirectory ( new java . io . File ( com . microsoft . azure . management . appservice . samples . ManageWebAppSourceControl . class . | org . junit . Assert . assertTrue ( com . microsoft . azure . management . appservice . samples . ManageWebAppSourceControl . runSample ( azure ) ) |
testJmxRegistration ( ) { final javax . management . ObjectName oname = gkoPool . getJmxName ( ) ; final javax . management . MBeanServer mbs = java . lang . management . ManagementFactory . getPlatformMBeanServer ( ) ; final java . util . Set < javax . management . ObjectName > result = mbs . queryNames ( oname , null ) ; "<AssertPlaceHolder>" ; } size ( ) { lock . lock ( ) ; try { return count ; } finally { lock . unlock ( ) ; } } | org . junit . Assert . assertEquals ( 1 , result . size ( ) ) |
testDecodeScopes ( ) { org . apache . oltu . oauth2 . common . utils . Set < java . lang . String > expected = new org . apache . oltu . oauth2 . common . utils . HashSet < java . lang . String > ( ) ; expected . add ( "email" ) ; expected . add ( "full_profile" ) ; org . apache . oltu . oauth2 . common . utils . Set < java . lang . String > scopes = org . apache . oltu . oauth2 . common . utils . OAuthUtils . decodeScopes ( "email<sp>full_profile" ) ; "<AssertPlaceHolder>" ; } decodeScopes ( java . lang . String ) { java . util . Set < java . lang . String > scopes = new java . util . HashSet < java . lang . String > ( ) ; if ( ! ( org . apache . oltu . oauth2 . common . utils . OAuthUtils . isEmpty ( s ) ) ) { java . util . StringTokenizer tokenizer = new java . util . StringTokenizer ( s , "<sp>" ) ; while ( tokenizer . hasMoreElements ( ) ) { scopes . add ( tokenizer . nextToken ( ) ) ; } } return scopes ; } | org . junit . Assert . assertEquals ( expected , scopes ) |
testJobWithSimpleJobIdentifier ( ) { java . lang . String testJob = "test" ; jobInstance = jobInstanceDao . createJobInstance ( testJob , jobParameters ) ; com . mongodb . DBObject dbObject = db . getCollection ( org . springframework . batch . mongo . dao . JobInstance . class . getSimpleName ( ) ) . findOne ( new com . mongodb . BasicDBObject ( org . springframework . batch . mongo . dao . MongoJobInstanceDao . JOB_INSTANCE_ID_KEY , jobInstance . getId ( ) ) ) ; "<AssertPlaceHolder>" ; } createJobInstance ( java . lang . String , org . springframework . batch . core . JobParameters ) { org . springframework . util . Assert . notNull ( jobName , "Job<sp>name<sp>must<sp>not<sp>be<sp>null." ) ; org . springframework . util . Assert . notNull ( jobParameters , "JobParameters<sp>must<sp>not<sp>be<sp>null." ) ; org . springframework . util . Assert . state ( ( ( getJobInstance ( jobName , jobParameters ) ) == null ) , "JobInstance<sp>must<sp>not<sp>already<sp>exist" ) ; java . lang . Long jobId = getNextId ( org . springframework . batch . core . JobInstance . class . getSimpleName ( ) ) ; org . springframework . batch . core . JobInstance jobInstance = new org . springframework . batch . core . JobInstance ( jobId , jobParameters , jobName ) ; jobInstance . incrementVersion ( ) ; org . springframework . batch . mongo . dao . Map < java . lang . String , org . springframework . batch . core . JobParameter > jobParams = jobParameters . getParameters ( ) ; org . springframework . batch . mongo . dao . Map < java . lang . String , java . lang . Object > paramMap = new org . springframework . batch . mongo . dao . HashMap < java . lang . String , java . lang . Object > ( jobParams . size ( ) ) ; for ( Map . Entry < java . lang . String , org . springframework . batch . core . JobParameter > entry : jobParams . entrySet ( ) ) { paramMap . put ( entry . getKey ( ) . replaceAll ( org . springframework . batch . mongo . dao . DOT_STRING , org . springframework . batch . mongo . dao . DOT_ESCAPE_STRING ) , entry . getValue ( ) . getValue ( ) ) ; } getCollection ( ) . save ( start ( ) . add ( org . springframework . batch . mongo . dao . MongoJobInstanceDao . JOB_INSTANCE_ID_KEY , jobId ) . add ( org . springframework . batch . mongo . dao . MongoJobInstanceDao . JOB_NAME_KEY , jobName ) . add ( org . springframework . batch . mongo . dao . MongoJobInstanceDao . JOB_KEY_KEY , createJobKey ( jobParameters ) ) . add ( org . springframework . batch . mongo . dao . VERSION_KEY , jobInstance . getVersion ( ) ) . add ( org . springframework . batch . mongo . dao . MongoJobInstanceDao . JOB_PARAMETERS_KEY , new com . mongodb . BasicDBObject ( paramMap ) ) . get ( ) ) ; return jobInstance ; } | org . junit . Assert . assertEquals ( "test" , dbObject . get ( org . springframework . batch . mongo . dao . MongoJobInstanceDao . JOB_NAME_KEY ) ) |
test_GetItem_By_AutomationId ( ) { when ( element . findFirst ( mmarquee . automation . BaseAutomationTest . isTreeScope ( TreeScope . Descendants ) , any ( ) ) ) . thenReturn ( targetElement ) ; mmarquee . automation . controls . AutomationList list = new mmarquee . automation . controls . AutomationList ( new mmarquee . automation . controls . ElementBuilder ( element ) . addPattern ( selection ) ) ; mmarquee . automation . controls . AutomationListItem item = list . getItemByAutomationId ( "myID" ) ; "<AssertPlaceHolder>" ; verify ( element , atLeastOnce ( ) ) . findFirst ( any ( ) , any ( ) ) ; } getElement ( ) { return this . element ; } | org . junit . Assert . assertEquals ( targetElement , item . getElement ( ) ) |
testExistZipFullPublish ( ) { setFunctionFound ( true ) ; setPublishVersion ( "1" ) ; com . xti . jenkins . plugin . awslambda . upload . DeployConfig deployConfig = getDeployConfig ( ) ; deployConfig . setPublish ( true ) ; java . lang . Boolean result = lambdaDeployService . deployLambda ( deployConfig , getFunctionCode ( ) , UpdateModeValue . Full ) ; calledGetFunction ( ) ; calledCreateFunction ( false ) ; calledCreateAlias ( false ) ; calledUpdateCode ( true ) ; calledUpdateConfiguration ( true ) ; "<AssertPlaceHolder>" ; } calledUpdateConfiguration ( java . lang . Boolean ) { if ( called ) { org . mockito . ArgumentCaptor < com . amazonaws . services . lambda . model . UpdateFunctionConfigurationRequest > args = org . mockito . ArgumentCaptor . forClass ( com . amazonaws . services . lambda . model . UpdateFunctionConfigurationRequest . class ) ; verify ( awsLambdaClient , times ( 1 ) ) . updateFunctionConfiguration ( args . capture ( ) ) ; com . amazonaws . services . lambda . model . UpdateFunctionConfigurationRequest expected = new com . amazonaws . services . lambda . model . UpdateFunctionConfigurationRequest ( ) . withDescription ( description ) . withFunctionName ( functionName ) . withHandler ( handler ) . withMemorySize ( memory ) . withRole ( role ) . withRuntime ( runtime ) . withVpcConfig ( new com . amazonaws . services . lambda . model . VpcConfig ( ) . withSubnetIds ( subnets ) . withSecurityGroupIds ( securityGroups ) ) . withTimeout ( timeout ) . withKMSKeyArn ( kmsArn ) . withEnvironment ( new com . amazonaws . services . lambda . model . Environment ( ) . withVariables ( environment ) ) . withDeadLetterConfig ( new com . amazonaws . services . lambda . model . DeadLetterConfig ( ) . withTargetArn ( deadLetterQueueArn ) ) ; org . junit . Assert . assertEquals ( expected , args . getValue ( ) ) ; } else { verify ( awsLambdaClient , never ( ) ) . updateFunctionConfiguration ( any ( com . amazonaws . services . lambda . model . UpdateFunctionConfigurationRequest . class ) ) ; } } | org . junit . Assert . assertTrue ( result ) |
testPathWithSpaces ( ) { org . apache . hadoop . tools . TestHadoopArchives . createFile ( inputPath , fs , "c<sp>c" ) ; final org . apache . hadoop . fs . Path sub1 = new org . apache . hadoop . fs . Path ( inputPath , "sub<sp>1" ) ; fs . mkdirs ( sub1 ) ; org . apache . hadoop . tools . TestHadoopArchives . createFile ( sub1 , fs , "file<sp>x<sp>y<sp>z" ) ; org . apache . hadoop . tools . TestHadoopArchives . createFile ( sub1 , fs , "file" ) ; org . apache . hadoop . tools . TestHadoopArchives . createFile ( sub1 , fs , "x" ) ; org . apache . hadoop . tools . TestHadoopArchives . createFile ( sub1 , fs , "y" ) ; org . apache . hadoop . tools . TestHadoopArchives . createFile ( sub1 , fs , "z" ) ; final org . apache . hadoop . fs . Path sub2 = new org . apache . hadoop . fs . Path ( inputPath , "sub<sp>1<sp>with<sp>suffix" ) ; fs . mkdirs ( sub2 ) ; org . apache . hadoop . tools . TestHadoopArchives . createFile ( sub2 , fs , "z" ) ; final org . apache . hadoop . fs . FsShell shell = new org . apache . hadoop . fs . FsShell ( conf ) ; final java . lang . String inputPathStr = inputPath . toUri ( ) . getPath ( ) ; final java . util . List < java . lang . String > originalPaths = org . apache . hadoop . tools . TestHadoopArchives . lsr ( shell , inputPathStr ) ; final java . lang . String fullHarPathStr = makeArchive ( ) ; final java . util . List < java . lang . String > harPaths = org . apache . hadoop . tools . TestHadoopArchives . lsr ( shell , fullHarPathStr ) ; "<AssertPlaceHolder>" ; } lsr ( org . apache . hadoop . fs . FsShell , java . lang . String ) { return org . apache . hadoop . tools . TestHadoopArchives . lsr ( shell , rootDir , null ) ; } | org . junit . Assert . assertEquals ( originalPaths , harPaths ) |
testRight ( ) { java . lang . String position = net . designforcode . controllers . BooksViewed . determineBookPosition ( 1 ) ; "<AssertPlaceHolder>" ; } determineBookPosition ( java . lang . Integer ) { return position == 0 ? "leftBook" : "rightBook" ; } | org . junit . Assert . assertThat ( position , org . hamcrest . CoreMatchers . is ( "rightBook" ) ) |
testGetEpochTime ( ) { org . jinstagram . realtime . SubscriptionResponseObject fixture = new org . jinstagram . realtime . SubscriptionResponseObject ( ) ; fixture . setObjectId ( "" ) ; fixture . setEpochTime ( 1L ) ; fixture . setChangedAspect ( "" ) ; fixture . setObject ( "" ) ; fixture . setSubscriptionId ( "" ) ; long result = fixture . getEpochTime ( ) ; "<AssertPlaceHolder>" ; } getEpochTime ( ) { return epochTime ; } | org . junit . Assert . assertEquals ( 1L , result ) |
verifyThatAllPropertiesInSourceOverrideTargetProperties ( ) { java . util . Properties source = new java . util . Properties ( ) ; java . util . Properties target = new java . util . Properties ( ) ; source . setProperty ( "key-1" , "some-value" ) ; target . setProperty ( "key-1" , "some-other-value" ) ; me . magnet . consultant . PropertiesUtil . sync ( source , target ) ; "<AssertPlaceHolder>" ; } sync ( java . util . Properties , java . util . Properties ) { checkNotNull ( source , "You<sp>must<sp>specify<sp>a<sp>'source'<sp>Properties<sp>object!" ) ; checkNotNull ( target , "You<sp>must<sp>specify<sp>a<sp>'source'<sp>Properties<sp>object!" ) ; java . util . Set < java . lang . String > sourceKeys = source . stringPropertyNames ( ) ; java . util . Set < java . lang . String > targetKeys = target . stringPropertyNames ( ) ; java . util . Set < java . lang . String > added = com . google . common . collect . Sets . newHashSet ( com . google . common . collect . Sets . difference ( sourceKeys , targetKeys ) ) ; java . util . Set < java . lang . String > modified = com . google . common . collect . Sets . newHashSet ( com . google . common . collect . Sets . intersection ( sourceKeys , targetKeys ) ) ; java . util . Set < java . lang . String > removed = com . google . common . collect . Sets . newHashSet ( com . google . common . collect . Sets . difference ( targetKeys , sourceKeys ) ) ; java . util . Map < java . lang . String , org . apache . commons . lang3 . tuple . Pair < java . lang . String , java . lang . String > > changes = com . google . common . collect . Maps . newHashMap ( ) ; added . forEach ( ( key ) -> { java . lang . String newValue = source . getProperty ( key ) ; changes . put ( key , org . apache . commons . lang3 . tuple . Pair . of ( null , newValue ) ) ; target . setProperty ( key , newValue ) ; } ) ; modified . forEach ( ( key ) -> { java . lang . String oldValue = target . getProperty ( key ) ; java . lang . String newValue = source . getProperty ( key ) ; changes . put ( key , org . apache . commons . lang3 . tuple . Pair . of ( oldValue , newValue ) ) ; target . setProperty ( key , newValue ) ; } ) ; removed . forEach ( ( key ) -> { java . lang . String oldValue = target . getProperty ( key ) ; changes . put ( key , org . apache . commons . lang3 . tuple . Pair . of ( oldValue , null ) ) ; target . remove ( key ) ; } ) ; return changes ; } | org . junit . Assert . assertEquals ( "some-value" , target . getProperty ( "key-1" ) ) |
testContentsExcluded ( ) { ds . setBasedir ( new java . io . File ( "." ) ) ; ds . setIncludes ( new java . lang . String [ ] { "**" } ) ; ds . addDefaultExcludes ( ) ; ds . ensureNonPatternSetsReady ( ) ; java . io . File f = new java . io . File ( ".svn" ) ; org . apache . tools . ant . types . selectors . TokenizedPath p = new org . apache . tools . ant . types . selectors . TokenizedPath ( f . getAbsolutePath ( ) ) ; "<AssertPlaceHolder>" ; } contentsExcluded ( org . apache . tools . ant . types . selectors . TokenizedPath ) { return java . util . stream . Stream . of ( excludePatterns ) . filter ( ( p ) -> p . endsWith ( SelectorUtils . DEEP_TREE_MATCH ) ) . map ( TokenizedPattern :: withoutLastToken ) . anyMatch ( ( wlt ) -> wlt . matchPath ( path , isCaseSensitive ( ) ) ) ; } | org . junit . Assert . assertTrue ( ds . contentsExcluded ( p ) ) |
testSetSecondInterpolationValue ( ) { org . orbisgis . coremap . renderer . se . Style st = getStyle ( org . orbisgis . legend . analyzer . PROP_LINE ) ; org . orbisgis . coremap . renderer . se . LineSymbolizer ls = ( ( org . orbisgis . coremap . renderer . se . LineSymbolizer ) ( st . getRules ( ) . get ( 0 ) . getCompositeSymbolizer ( ) . getSymbolizerList ( ) . get ( 0 ) ) ) ; org . orbisgis . legend . thematic . proportional . ProportionalLine usl = new org . orbisgis . legend . thematic . proportional . ProportionalLine ( ls ) ; usl . setSecondValue ( 24.3 ) ; "<AssertPlaceHolder>" ; } getSecondValue ( ) { return interp . getInterpolationPoint ( 1 ) . getValue ( ) . getValue ( null , 0 ) ; } | org . junit . Assert . assertTrue ( ( ( usl . getSecondValue ( ) ) == 24.3 ) ) |
forkingContextDoesNotCarryDeadline ( ) { io . grpc . Deadline deadline = io . grpc . Deadline . after ( 1 , TimeUnit . HOURS ) ; io . grpc . Context . CancellableContext parent = Context . ROOT . withDeadline ( deadline , scheduler ) ; io . grpc . Context fork = parent . fork ( ) ; "<AssertPlaceHolder>" ; } getDeadline ( ) { return deadline ; } | org . junit . Assert . assertNull ( fork . getDeadline ( ) ) |
compareToInputNotNullOutputPositive2 ( ) { final com . freedomotic . environment . GraphEdge objectUnderTest = new com . freedomotic . environment . GraphEdge ( null , ( - 1 ) , 2147483323 ) ; final com . freedomotic . environment . GraphEdge other = new com . freedomotic . environment . GraphEdge ( null , ( - 1 ) , 1879047867 ) ; final int retval = objectUnderTest . compareTo ( other ) ; "<AssertPlaceHolder>" ; } compareTo ( com . freedomotic . i18n . ComboLanguage ) { return description . compareTo ( other . toString ( ) ) ; } | org . junit . Assert . assertEquals ( 1 , retval ) |
testGetAllRefreshTokens ( ) { java . util . Set < org . mitre . oauth2 . model . OAuth2RefreshTokenEntity > tokens = repository . getAllRefreshTokens ( ) ; "<AssertPlaceHolder>" ; } getAllRefreshTokens ( ) { javax . persistence . TypedQuery < org . mitre . oauth2 . model . OAuth2RefreshTokenEntity > query = manager . createNamedQuery ( OAuth2RefreshTokenEntity . QUERY_ALL , org . mitre . oauth2 . model . OAuth2RefreshTokenEntity . class ) ; return new java . util . LinkedHashSet ( query . getResultList ( ) ) ; } | org . junit . Assert . assertEquals ( 5 , tokens . size ( ) ) |
isEndElement ( ) { when ( this . reader . isEndElement ( ) ) . thenReturn ( true ) ; boolean result = this . filter . isEndElement ( ) ; "<AssertPlaceHolder>" ; } isEndElement ( ) { when ( this . reader . isEndElement ( ) ) . thenReturn ( true ) ; boolean result = this . filter . isEndElement ( ) ; org . junit . Assert . assertTrue ( result ) ; } | org . junit . Assert . assertTrue ( result ) |
ttestUpdateGroupAttributes_noEntityManager_assertException ( ) { final org . nhindirect . config . store . dao . impl . CertPolicyDaoImpl dao = new org . nhindirect . config . store . dao . impl . CertPolicyDaoImpl ( ) ; boolean exceptionOccured = false ; try { dao . updateGroupAttributes ( 12345 , "" ) ; } catch ( java . lang . IllegalStateException ex ) { exceptionOccured = true ; } "<AssertPlaceHolder>" ; } | org . junit . Assert . assertTrue ( exceptionOccured ) |
testParseMaxKeyInArrayStrict ( ) { final com . allanbank . mongodb . bson . json . JsonParser parser = new com . allanbank . mongodb . bson . json . JsonParser ( ) ; final java . lang . Object doc = parser . parse ( "{<sp>a<sp>:<sp>[{<sp>$maxKey:1}]<sp>}" ) ; final com . allanbank . mongodb . bson . builder . DocumentBuilder b = com . allanbank . mongodb . bson . builder . BuilderFactory . start ( ) ; b . pushArray ( "a" ) . addMaxKey ( ) ; "<AssertPlaceHolder>" ; } build ( ) { final java . util . List < com . allanbank . mongodb . bson . Element > elements = subElements ( ) ; return elements . toArray ( new com . allanbank . mongodb . bson . Element [ elements . size ( ) ] ) ; } | org . junit . Assert . assertEquals ( b . build ( ) , doc ) |
convertStringArrayA_returnArray ( ) { java . lang . String [ ] expected = new java . lang . String [ ] { "THIS_FIELD<sp>:<sp>field_this<sp>" , "<sp>THAT_FIELD<sp>:<sp>field_that" } ; java . lang . String string = "<sp>THIS_FIELD<sp>:<sp>field_this<sp>,<sp>THAT_FIELD<sp>:<sp>field_that<sp>" ; java . lang . String [ ] actual = dk . i2m . converge . plugins . drupal . DrupalUtils . convertStringArrayA ( string ) ; "<AssertPlaceHolder>" ; } convertStringArrayA ( java . lang . String ) { return string . trim ( ) . split ( dk . i2m . converge . plugins . drupal . DrupalUtils . SEPARATOR_A ) ; } | org . junit . Assert . assertArrayEquals ( expected , actual ) |
testSingle ( ) { java . util . List < java . lang . Boolean > result = com . squareup . cascading_helpers . TestHelpers . exec ( new com . squareup . cascading_helpers . filter . FilterEquals ( 1 ) , new cascading . tuple . Fields ( "blah" ) , new cascading . tuple . Tuple ( 1 ) , new cascading . tuple . Tuple ( 2 ) ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( java . util . Arrays . asList ( true , false ) , result ) |
test_GetProgresBar_By_AutomationId ( ) { when ( element . findFirst ( mmarquee . automation . BaseAutomationTest . isTreeScope ( TreeScope . Descendants ) , any ( ) ) ) . thenReturn ( targetElement ) ; mmarquee . automation . controls . AutomationProgressBar bar = spyWndw . getProgressBar ( mmarquee . automation . controls . Search . getBuilder ( ) . automationId ( "myID" ) . build ( ) ) ; "<AssertPlaceHolder>" ; verify ( spyWndw ) . createAutomationIdPropertyCondition ( "myID" ) ; verify ( spyWndw ) . createControlTypeCondition ( ControlType . ProgressBar ) ; verify ( element , atLeastOnce ( ) ) . findFirst ( any ( ) , any ( ) ) ; } getElement ( ) { return this . element ; } | org . junit . Assert . assertEquals ( targetElement , bar . getElement ( ) ) |
testArrayConstructorReferenceBoxed ( ) { org . robovm . rt . lambdas . Java8Test . ArrayCtorBoxed ctor = org . robovm . rt . lambdas . Java8Test . ArrayElem [ ] [ ] [ ] :: new ; org . robovm . rt . lambdas . Java8Test . ArrayElem [ ] [ ] [ ] array = ctor . copy ( 100 ) ; "<AssertPlaceHolder>" ; } copy ( java . nio . CharSequenceAdapter ) { java . nio . CharSequenceAdapter buf = new java . nio . CharSequenceAdapter ( other . sequence ) ; buf . limit = other . limit ; buf . position = other . position ; buf . mark = other . mark ; return buf ; } | org . junit . Assert . assertEquals ( 100 , array . length ) |
testPublicTestVectors ( ) { for ( com . google . crypto . tink . aead . AesEaxKeyManagerTest . PublicTestVector t : publicTestVectors ) { if ( com . google . crypto . tink . TestUtil . shouldSkipTestWithAesKeySize ( t . keyValue . length ) ) { continue ; } com . google . crypto . tink . Aead aead = getRawAesEax ( t . keyValue , t . iv . length ) ; try { byte [ ] ciphertext = com . google . crypto . tink . subtle . Bytes . concat ( t . iv , t . ciphertext , t . tag ) ; byte [ ] plaintext = aead . decrypt ( ciphertext , t . aad ) ; "<AssertPlaceHolder>" ; } catch ( java . security . GeneralSecurityException e ) { org . junit . Assert . fail ( ( ( ( "Should<sp>not<sp>fail<sp>at<sp>" + ( t . name ) ) + ",<sp>but<sp>thrown<sp>exception<sp>" ) + e ) ) ; } } } decrypt ( byte [ ] , byte [ ] ) { try { java . nio . ByteBuffer buffer = java . nio . ByteBuffer . wrap ( ciphertext ) ; int encryptedDekSize = buffer . getInt ( ) ; if ( ( encryptedDekSize <= 0 ) || ( encryptedDekSize > ( ( ciphertext . length ) - ( com . google . crypto . tink . aead . KmsEnvelopeAead . LENGTH_ENCRYPTED_DEK ) ) ) ) { throw new java . security . GeneralSecurityException ( "invalid<sp>ciphertext" ) ; } byte [ ] encryptedDek = new byte [ encryptedDekSize ] ; buffer . get ( encryptedDek , 0 , encryptedDekSize ) ; byte [ ] payload = new byte [ buffer . remaining ( ) ] ; buffer . get ( payload , 0 , buffer . remaining ( ) ) ; byte [ ] dek = remote . decrypt ( encryptedDek , com . google . crypto . tink . aead . KmsEnvelopeAead . EMPTY_AAD ) ; com . google . crypto . tink . Aead aead = com . google . crypto . tink . Registry . getPrimitive ( dekTemplate . getTypeUrl ( ) , dek , com . google . crypto . tink . Aead . class ) ; return aead . decrypt ( payload , associatedData ) ; } catch ( java . lang . IndexOutOfBoundsException | java . nio . BufferUnderflowException | java . lang . NegativeArraySizeException e ) { throw new java . security . GeneralSecurityException ( "invalid<sp>ciphertext" , e ) ; } } | org . junit . Assert . assertArrayEquals ( plaintext , t . plaintext ) |
testUnlinkFailesOnFilesLookup ( ) { final com . allanbank . mongodb . bson . builder . DocumentBuilder fileResult = com . allanbank . mongodb . bson . builder . BuilderFactory . start ( ) ; fileResult . addString ( "_id" , "id" ) ; final com . allanbank . mongodb . bson . builder . DocumentBuilder chunkResult = com . allanbank . mongodb . bson . builder . BuilderFactory . start ( ) ; chunkResult . addBinary ( "data" , new byte [ ] { 1 , 2 , 3 , 4 } ) ; final com . allanbank . mongodb . MongoDatabase mockDb = createMock ( com . allanbank . mongodb . MongoDatabase . class ) ; final com . allanbank . mongodb . MongoCollection mockFiles = createMock ( com . allanbank . mongodb . MongoCollection . class ) ; final com . allanbank . mongodb . MongoCollection mockChunks = createMock ( com . allanbank . mongodb . MongoCollection . class ) ; expect ( mockDb . getCollection ( ( "fs" + ( GridFs . FILES_SUFFIX ) ) ) ) . andReturn ( mockFiles ) ; expect ( mockDb . getCollection ( ( "fs" + ( GridFs . CHUNKS_SUFFIX ) ) ) ) . andReturn ( mockChunks ) ; expect ( mockFiles . findOne ( anyObject ( com . allanbank . mongodb . bson . DocumentAssignable . class ) ) ) . andReturn ( null ) ; replay ( mockDb , mockFiles , mockChunks ) ; final com . allanbank . mongodb . gridfs . GridFs fs = new com . allanbank . mongodb . gridfs . GridFs ( mockDb ) ; "<AssertPlaceHolder>" ; verify ( mockDb , mockFiles , mockChunks ) ; } unlink ( com . allanbank . mongodb . bson . element . ObjectId ) { final com . allanbank . mongodb . bson . Document fileDoc = myFilesCollection . findOne ( where ( com . allanbank . mongodb . gridfs . GridFs . ID_FIELD ) . equals ( id ) ) ; if ( fileDoc == null ) { return false ; } return doUnlink ( fileDoc ) ; } | org . junit . Assert . assertFalse ( fs . unlink ( "foo" ) ) |
handle_shouldTrimWhitespaceFromName ( ) { java . lang . String NAME = "the<sp>location<sp>name" ; org . openmrs . api . handler . OpenmrsMetadataSaveHandler handler = new org . openmrs . api . handler . OpenmrsMetadataSaveHandler ( ) ; org . openmrs . Location loc = new org . openmrs . Location ( ) ; loc . setName ( ( ( "<sp>" + NAME ) + "<sp>" ) ) ; handler . handle ( loc , null , null , null ) ; "<AssertPlaceHolder>" ; } getName ( ) { if ( getNames ( ) . isEmpty ( ) ) { if ( org . openmrs . Concept . log . isDebugEnabled ( ) ) { org . openmrs . Concept . log . debug ( ( "there<sp>are<sp>no<sp>names<sp>defined<sp>for:<sp>" + ( conceptId ) ) ) ; } return null ; } for ( java . util . Locale currentLocale : org . openmrs . util . LocaleUtility . getLocalesInOrder ( ) ) { org . openmrs . ConceptName preferredName = getPreferredName ( currentLocale ) ; if ( preferredName != null ) { return preferredName ; } org . openmrs . ConceptName fullySpecifiedName = getFullySpecifiedName ( currentLocale ) ; if ( fullySpecifiedName != null ) { return fullySpecifiedName ; } if ( ( ! ( org . apache . commons . lang3 . StringUtils . isBlank ( currentLocale . getCountry ( ) ) ) ) || ( ! ( org . apache . commons . lang3 . StringUtils . isBlank ( currentLocale . getVariant ( ) ) ) ) ) { java . util . Locale broaderLocale = new java . util . Locale ( currentLocale . getLanguage ( ) ) ; org . openmrs . ConceptName prefNameInBroaderLoc = getPreferredName ( broaderLocale ) ; if ( prefNameInBroaderLoc != null ) { return prefNameInBroaderLoc ; } org . openmrs . ConceptName fullySpecNameInBroaderLoc = getFullySpecifiedName ( broaderLocale ) ; if ( fullySpecNameInBroaderLoc != null ) { return fullySpecNameInBroaderLoc ; } } } for ( org . openmrs . ConceptName cn : getNames ( ) ) { if ( cn . isFullySpecifiedName ( ) ) { return cn ; } } if ( ! ( getSynonyms ( ) . isEmpty ( ) ) ) { return getSynonyms ( ) . iterator ( ) . next ( ) ; } return null ; } | org . junit . Assert . assertEquals ( NAME , loc . getName ( ) ) |
getLocationCountShouldWork ( ) { when ( locationRepository . count ( ) ) . thenReturn ( 4711L ) ; final ac . simons . biking2 . tracker . LocationService locationService = new ac . simons . biking2 . tracker . LocationService ( locationRepository , simpMessagingTemplate ) ; "<AssertPlaceHolder>" ; } getLocationCount ( ) { return locationRepository . count ( ) ; } | org . junit . Assert . assertEquals ( 4711L , locationService . getLocationCount ( ) ) |
testMultipleEncodeReferenceCount ( ) { io . netty . channel . embedded . EmbeddedChannel channel = new io . netty . channel . embedded . EmbeddedChannel ( new io . netty . handler . codec . serialization . CompatibleObjectEncoder ( ) ) ; io . netty . handler . codec . serialization . CompatibleObjectEncoderTest . testEncode ( channel , new io . netty . handler . codec . serialization . CompatibleObjectEncoderTest . TestSerializable ( 6 , 8 ) ) ; io . netty . handler . codec . serialization . CompatibleObjectEncoderTest . testEncode ( channel , new io . netty . handler . codec . serialization . CompatibleObjectEncoderTest . TestSerializable ( 10 , 5 ) ) ; io . netty . handler . codec . serialization . CompatibleObjectEncoderTest . testEncode ( channel , new io . netty . handler . codec . serialization . CompatibleObjectEncoderTest . TestSerializable ( 1 , 5 ) ) ; "<AssertPlaceHolder>" ; } finishAndReleaseAll ( ) { return finish ( true ) ; } | org . junit . Assert . assertFalse ( channel . finishAndReleaseAll ( ) ) |
testGetInstance ( ) { "<AssertPlaceHolder>" ; } getInstance ( ) { return de . strullerbaumann . visualee . filter . boundary . FilterContainer . FilterContainerHolder . INSTANCE ; } | org . junit . Assert . assertNotNull ( de . strullerbaumann . visualee . filter . boundary . FilterContainer . getInstance ( ) ) |
getPotentialDuplicate ( ) { org . mockito . Mockito . when ( deduplicationService . getPotentialDuplicateByUid ( org . mockito . Mockito . eq ( "1" ) ) ) . thenReturn ( new org . hisp . dhis . deduplication . PotentialDuplicate ( "teiA" ) ) ; org . hisp . dhis . deduplication . PotentialDuplicate pd = controller . getPotentialDuplicate ( "1" ) ; "<AssertPlaceHolder>" ; } getTeiA ( ) { return teiA ; } | org . junit . Assert . assertEquals ( "teiA" , pd . getTeiA ( ) ) |
testBuildCategoryTree ( ) { org . jahia . services . categories . Category rootCategory = org . jahia . services . categories . Category . createCategory ( "firstRoot" , null ) ; org . jahia . services . categories . Category newCategory = org . jahia . services . categories . Category . createCategory ( "rootChild" , rootCategory ) ; buildCategoryTree ( newCategory , 4 , 3 ) ; int sizeOfTree = 0 ; for ( int i = 1 ; i <= 4 ; i ++ ) { sizeOfTree += java . lang . Math . pow ( 3 , i ) ; } final java . util . List < org . jahia . services . categories . Category > childCategories = newCategory . getChildCategories ( true ) ; "<AssertPlaceHolder>" ; deleteCategoryWithChildren ( rootCategory ) ; } size ( ) { return properties . keySet ( ) . size ( ) ; } | org . junit . Assert . assertTrue ( ( ( childCategories . size ( ) ) == sizeOfTree ) ) |
whenOfCalled_thenCreateEmptyImmutableMap ( ) { com . google . common . collect . ClassToInstanceMap < org . baeldung . guava . Action > map = com . google . common . collect . ImmutableClassToInstanceMap . of ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( total ) == 0 ; } | org . junit . Assert . assertTrue ( map . isEmpty ( ) ) |
testHColumnValidName ( ) { boolean exceptionThrown ; try { new org . apache . hadoop . hbase . HColumnDescriptor ( "\\test\\abc" ) ; } catch ( java . lang . IllegalArgumentException iae ) { exceptionThrown = true ; "<AssertPlaceHolder>" ; } } | org . junit . Assert . assertTrue ( exceptionThrown ) |
escapeQueryAttributes_should_work_without_item_id ( ) { final java . lang . String convertedQuery = org . springframework . data . simpledb . query . QueryUtils . escapeQueryAttributes ( "select<sp>*<sp>from<sp>customers" , "customer_id" ) ; "<AssertPlaceHolder>" ; } escapeQueryAttributes ( java . lang . String , java . lang . String ) { java . lang . String escapedQuery = rawQuery . replaceAll ( ( ( "\\s" + idFieldName ) + "\\s" ) , "<sp>itemName()<sp>" ) ; if ( escapedQuery . endsWith ( idFieldName ) ) { escapedQuery = escapedQuery . substring ( 0 , ( ( escapedQuery . length ( ) ) - ( idFieldName . length ( ) ) ) ) ; escapedQuery += "itemName()" ; } return escapedQuery ; } | org . junit . Assert . assertThat ( convertedQuery , org . hamcrest . CoreMatchers . is ( "select<sp>*<sp>from<sp>customers" ) ) |
queryBeanX ( ) { org . n3r . eql . eqler . MyEqlerBean bean = org . n3r . eql . eqler . MyEqlerTest . myEqler . queryBeanX ( "bb" ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertThat ( bean , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( new org . n3r . eql . eqler . MyEqlerBean ( "bb" ) ) ) ) |
shouldNotChangeCellWhenEditingAndArrowRightKeyIsPressed ( ) { final com . vaadin . addon . spreadsheet . elements . SheetCellElement b2 = spreadSheet . getCellAt ( "B2" ) ; b2 . setValue ( "123" ) ; sheetController . selectCell ( "B2" ) ; new org . openqa . selenium . interactions . Actions ( getDriver ( ) ) . sendKeys ( Keys . F2 ) . build ( ) . perform ( ) ; new org . openqa . selenium . interactions . Actions ( getDriver ( ) ) . sendKeys ( Keys . ARROW_RIGHT ) . build ( ) . perform ( ) ; java . lang . String selectedCell = sheetController . getSelectedCell ( ) ; "<AssertPlaceHolder>" ; } getSelectedCell ( ) { java . lang . String elemClass = driver . findElement ( org . openqa . selenium . By . cssSelector ( ".sheet-selection" ) ) . getAttribute ( "class" ) ; int rowStart = elemClass . indexOf ( "row" ) ; if ( rowStart == ( - 1 ) ) { return "A1" ; } int k = rowStart + ( "row" . length ( ) ) ; java . lang . String rowNumber = "" ; while ( k < ( elemClass . length ( ) ) ) { char digit = elemClass . charAt ( k ) ; if ( digit == '<sp>' ) { break ; } rowNumber += elemClass . charAt ( k ) ; k ++ ; } int colStart = elemClass . indexOf ( "col" ) ; k = colStart + ( "col" . length ( ) ) ; java . lang . String colNumberStr = "" ; while ( k < ( elemClass . length ( ) ) ) { char digit = elemClass . charAt ( k ) ; if ( digit == '<sp>' ) { break ; } colNumberStr += elemClass . charAt ( k ) ; k ++ ; } int colNumber = java . lang . Integer . parseInt ( colNumberStr ) ; int dividend = colNumber ; java . lang . String columnName = "" ; int modulo ; while ( dividend > 0 ) { modulo = ( dividend - 1 ) % 26 ; columnName = ( ( char ) ( 65 + modulo ) ) + columnName ; dividend = ( dividend - modulo ) / 26 ; } return columnName + rowNumber ; } | org . junit . Assert . assertThat ( selectedCell , org . hamcrest . CoreMatchers . is ( "B2" ) ) |
testOneInstance ( ) { net . roboconf . core . model . beans . Graphs graphs = buildGraphs ( ) ; net . roboconf . core . model . beans . Component vmComponent = net . roboconf . core . model . helpers . ComponentHelpers . findComponent ( graphs , "VM" ) ; "<AssertPlaceHolder>" ; net . roboconf . core . model . beans . Instance inst = new net . roboconf . core . model . beans . Instance ( "inst" ) . component ( vmComponent ) ; compareInstances ( graphs , java . util . Arrays . asList ( inst ) , false , false ) ; } findComponent ( net . roboconf . core . model . beans . Graphs , java . lang . String ) { net . roboconf . core . model . beans . Component result = null ; for ( java . util . Iterator < net . roboconf . core . model . beans . Component > it = net . roboconf . core . model . helpers . ComponentHelpers . findAllComponents ( graphs ) . iterator ( ) ; ( it . hasNext ( ) ) && ( result == null ) ; ) { net . roboconf . core . model . beans . Component c = it . next ( ) ; if ( name . equals ( c . getName ( ) ) ) result = c ; } return result ; } | org . junit . Assert . assertNotNull ( vmComponent ) |
testMaximumTimeMillisecondsViaSetter ( ) { final java . util . Random random = new java . util . Random ( java . lang . System . currentTimeMillis ( ) ) ; final com . allanbank . mongodb . builder . Distinct . Builder b = new com . allanbank . mongodb . builder . Distinct . Builder ( ) ; b . key ( "foo" ) ; final long value = random . nextLong ( ) ; b . setMaximumTimeMilliseconds ( value ) ; final com . allanbank . mongodb . builder . Distinct command = b . build ( ) ; "<AssertPlaceHolder>" ; } getMaximumTimeMilliseconds ( ) { return myMaximumTimeMilliseconds ; } | org . junit . Assert . assertThat ( command . getMaximumTimeMilliseconds ( ) , org . hamcrest . Matchers . is ( value ) ) |
getAGivenUnaccessibleGroupInAnotherDomainWithSemiDomainIsolation ( ) { try { com . stratelia . webactiv . util . GeneralPropertiesManagerHelper . setDomainVisibility ( DomainProperties . DVIS_ONE ) ; com . stratelia . webactiv . beans . admin . Group actualGroup = getTestResources ( ) . getAGroupNotInAnInternalDomain ( ) ; currentUser . setDomainId ( ( ( actualGroup . getDomainId ( ) ) + "0" ) ) ; java . lang . String path = buildURIPathOf ( actualGroup ) ; getAt ( path , com . silverpeas . profile . web . UserGroupProfileEntity . class ) ; org . junit . Assert . fail ( "The<sp>group<sp>shouldn't<sp>be<sp>get<sp>as<sp>it<sp>is<sp>unaccessible" ) ; } catch ( com . sun . jersey . api . client . UniformInterfaceException ex ) { int receivedStatus = ex . getResponse ( ) . getStatus ( ) ; int forbidden = Response . Status . FORBIDDEN . getStatusCode ( ) ; "<AssertPlaceHolder>" ; } } is ( T ) { return java . util . Objects . equals ( this . value , value ) ; } | org . junit . Assert . assertThat ( receivedStatus , org . hamcrest . Matchers . is ( forbidden ) ) |
shouldDeleteFlow ( ) { org . openkilda . model . Flow flow = buildTestFlow ( org . openkilda . persistence . repositories . impl . Neo4jFlowRepositoryTest . TEST_FLOW_ID , switchA , switchB ) ; org . openkilda . persistence . repositories . impl . Neo4jFlowRepositoryTest . flowRepository . createOrUpdate ( flow ) ; org . openkilda . persistence . repositories . impl . Neo4jFlowRepositoryTest . flowRepository . delete ( flow ) ; "<AssertPlaceHolder>" ; } findAll ( ) { return super . findAll ( ) . stream ( ) . map ( this :: completeWithPaths ) . collect ( java . util . stream . Collectors . toList ( ) ) ; } | org . junit . Assert . assertEquals ( 0 , org . openkilda . persistence . repositories . impl . Neo4jFlowRepositoryTest . flowRepository . findAll ( ) . size ( ) ) |
validateHeaderWarnings ( ) { javax . xml . transform . Source testXml = new javax . xml . transform . stream . StreamSource ( new org . springframework . core . io . ClassPathResource ( sample_wrong ) . getInputStream ( ) ) ; params . put ( "phase" , "warnings" ) ; try { schematron . validate ( testXml , new org . openehealth . ipf . commons . xml . SchematronProfile ( "schematron/cda_phmr/templates/2.16.840.1.113883.10.20.9.sch" , params ) ) ; org . junit . Assert . fail ( ) ; } catch ( org . openehealth . ipf . commons . core . modules . api . ValidationException ex ) { "<AssertPlaceHolder>" ; } } getCauses ( ) { return causes ; } | org . junit . Assert . assertEquals ( 4 , ex . getCauses ( ) . length ) |
testRequestWithInvalidNegativeLimitQuery ( ) { org . apache . hadoop . yarn . api . protocolrecords . GetApplicationsRequest request = org . apache . hadoop . yarn . server . resourcemanager . webapp . ApplicationsRequestBuilder . create ( ) . withLimit ( "-10" ) . build ( ) ; org . apache . hadoop . yarn . api . protocolrecords . GetApplicationsRequest expectedRequest = getDefaultRequest ( ) ; "<AssertPlaceHolder>" ; } getDefaultRequest ( ) { org . apache . hadoop . yarn . api . protocolrecords . GetApplicationsRequest req = org . apache . hadoop . yarn . api . protocolrecords . GetApplicationsRequest . newInstance ( ) ; req . setStartRange ( 0 , Long . MAX_VALUE ) ; req . setFinishRange ( 0 , Long . MAX_VALUE ) ; return req ; } | org . junit . Assert . assertEquals ( expectedRequest , request ) |
test ( ) { System . out . println ( "15" ) ; int count = 100000 ; java . util . concurrent . atomic . AtomicInteger failCount = new java . util . concurrent . atomic . AtomicInteger ( ) ; java . util . concurrent . CountDownLatch down = new java . util . concurrent . CountDownLatch ( count ) ; org . yx . rpc . client . Rpc . call ( "a.b.repeat" , "" ) ; long begin = java . lang . System . currentTimeMillis ( ) ; for ( int i = 0 ; i < count ; i ++ ) { java . lang . String msg = "asdf-" + i ; org . yx . rpc . client . Rpc . create ( "a.b.repeat" ) . paramInArray ( msg ) . timeout ( 30000 ) . callback ( ( result ) -> { if ( ! ( msg . equals ( result . optResult ( . class ) ) ) ) { failCount . incrementAndGet ( ) ; } down . countDown ( ) ; } ) . execute ( ) ; } down . await ( ) ; long time = ( java . lang . System . currentTimeMillis ( ) ) - begin ; System . out . println ( ( ( ( "" + time ) + "," ) + ( ( count * 1000.0 ) / time ) ) ) ; "<AssertPlaceHolder>" ; } get ( ) { return org . yx . rpc . client . route . HostChecker . holder ; } | org . junit . Assert . assertEquals ( 0 , failCount . get ( ) ) |
testStop ( ) { final org . apache . james . mime4j . parser . MimeStreamParser parser = new org . apache . james . mime4j . parser . MimeStreamParser ( ) ; org . apache . james . mime4j . parser . TestHandler handler = new org . apache . james . mime4j . parser . TestHandler ( ) { @ org . apache . james . mime4j . parser . Override public void endHeader ( ) { super . endHeader ( ) ; parser . stop ( ) ; } } ; parser . setContentHandler ( handler ) ; java . lang . String msg = "Subject:<sp>Yada<sp>yada" 1 + ( ( ( "From:<sp>foo@bar.com\r\n" + "\r\n" ) + "Line<sp>1\r\n" ) + "Line<sp>2\r\n" ) ; java . lang . String expected = "<message>\r\n" + ( ( ( ( ( ( ( ( ( ( "Subject:<sp>Yada<sp>yada" 4 + "Subject:<sp>Yada<sp>yada" 2 ) + "Subject:<sp>Yada<sp>yada" ) + "</field>\r\n" ) + "Subject:<sp>Yada<sp>yada" 2 ) + "From:<sp>foo@bar.com" ) + "</field>\r\n" ) + "</header>\r\n" ) + "Subject:<sp>Yada<sp>yada" 0 ) + "Subject:<sp>Yada<sp>yada" 3 ) + "</message>\r\n" ) ; parser . parse ( new java . io . ByteArrayInputStream ( msg . getBytes ( ) ) ) ; java . lang . String result = handler . sb . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return getAddress ( ) ; } | org . junit . Assert . assertEquals ( expected , result ) |
serializeDeserializeStreamingFinalFrame ( ) { "<AssertPlaceHolder>" ; } serializeDeserializeStreaming ( boolean ) { final com . amazonaws . encryptionsdk . model . CipherBlockHeaders cipherBlockHeaders = new com . amazonaws . encryptionsdk . model . CipherBlockHeaders ( nonce_ , sampleContentLen_ ) ; final byte [ ] headerBytes = cipherBlockHeaders . toByteArray ( ) ; final com . amazonaws . encryptionsdk . model . CipherBlockHeaders reconstructedHeaders = new com . amazonaws . encryptionsdk . model . CipherBlockHeaders ( ) ; reconstructedHeaders . setNonceLength ( ( ( short ) ( nonceLen_ ) ) ) ; int totalParsedBytes = 0 ; int bytesToParseLen = 1 ; int bytesParsed ; while ( ( reconstructedHeaders . isComplete ( ) ) == false ) { final byte [ ] bytesToParse = new byte [ bytesToParseLen ] ; java . lang . System . arraycopy ( headerBytes , totalParsedBytes , bytesToParse , 0 , bytesToParse . length ) ; bytesParsed = reconstructedHeaders . deserialize ( bytesToParse , 0 ) ; if ( bytesParsed == 0 ) { bytesToParseLen ++ ; } else { totalParsedBytes += bytesParsed ; bytesToParseLen = 1 ; } } final byte [ ] reconstructedHeaderBytes = reconstructedHeaders . toByteArray ( ) ; return java . util . Arrays . equals ( headerBytes , reconstructedHeaderBytes ) ? true : false ; } | org . junit . Assert . assertTrue ( serializeDeserializeStreaming ( true ) ) |
should_return_this ( ) { org . fest . assertions . api . ShortAssert returned = assertions . isNotGreaterThan ( expected ) ; "<AssertPlaceHolder>" ; } isNotGreaterThan ( byte ) { bytes . assertNotGreaterThan ( description , actual , expected ) ; return this ; } | org . junit . Assert . assertSame ( assertions , returned ) |
testBuild ( ) { org . lnu . is . domain . person . work . PersonWork context = new org . lnu . is . domain . person . work . PersonWork ( ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>PersonWork<sp>e<sp>WHERE<sp>e.status=:status<sp>AND<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . person . work . PersonWork > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setEntity ( T ) { this . entity = entity ; } | org . junit . Assert . assertEquals ( expectedQuery , actualQuery ) |
shouldApplyRelationshipTypeTokenCommandToTheStoreInRecovery ( ) { final org . neo4j . kernel . impl . api . BatchTransactionApplier applier = newApplier ( true ) ; final org . neo4j . kernel . impl . store . record . RelationshipTypeTokenRecord before = new org . neo4j . kernel . impl . store . record . RelationshipTypeTokenRecord ( 42 ) ; final org . neo4j . kernel . impl . store . record . RelationshipTypeTokenRecord after = new org . neo4j . kernel . impl . store . record . RelationshipTypeTokenRecord ( 42 ) ; after . setInUse ( true ) ; after . setNameId ( 323 ) ; final org . neo4j . kernel . impl . transaction . command . Command . RelationshipTypeTokenCommand command = new org . neo4j . kernel . impl . transaction . command . Command . RelationshipTypeTokenCommand ( before , after ) ; final org . neo4j . internal . kernel . api . NamedToken token = new org . neo4j . internal . kernel . api . NamedToken ( "token" , 21 ) ; when ( relationshipTypeTokenStore . getToken ( ( ( int ) ( command . getKey ( ) ) ) ) ) . thenReturn ( token ) ; boolean result = apply ( applier , command :: handle , transactionToApply ) ; "<AssertPlaceHolder>" ; verify ( relationshipTypeTokenStore , times ( 1 ) ) . setHighestPossibleIdInUse ( after . getId ( ) ) ; verify ( relationshipTypeTokenStore , times ( 1 ) ) . updateRecord ( after ) ; verify ( cacheAccess , times ( 1 ) ) . addRelationshipTypeToken ( token ) ; } apply ( org . neo4j . kernel . api . proc . Context , java . lang . Object [ ] , org . neo4j . kernel . api . ResourceTracker ) { try { java . lang . Thread . sleep ( 50 ) ; } catch ( java . lang . InterruptedException e ) { throw new org . neo4j . internal . kernel . api . exceptions . ProcedureException ( Status . General . UnknownError , e , "Interrupted" ) ; } return org . neo4j . collection . RawIterator . empty ( ) ; } | org . junit . Assert . assertFalse ( result ) |
testSearchByMultipleTypes ( ) { int pageSize = 10 ; org . codice . ddf . spatial . ogc . wfs . v2_0_0 . catalog . source . WfsSource source = getWfsSource ( org . codice . ddf . spatial . ogc . wfs . v2_0_0 . catalog . source . WfsSourceTest . ONE_TEXT_PROPERTY_SCHEMA , org . codice . ddf . spatial . ogc . wfs . v2_0_0 . catalog . source . MockWfsServer . getFilterCapabilities ( ) , GeospatialUtil . EPSG_4326_URN , 10 , false ) ; org . opengis . filter . Filter filter0 = builder . attribute ( Metacard . CONTENT_TYPE ) . is ( ) . equalTo ( ) . text ( ( ( org . codice . ddf . spatial . ogc . wfs . v2_0_0 . catalog . source . WfsSourceTest . SAMPLE_FEATURE_NAME ) + "8" ) ) ; org . opengis . filter . Filter filter1 = builder . attribute ( Metacard . CONTENT_TYPE ) . is ( ) . equalTo ( ) . text ( ( ( org . codice . ddf . spatial . ogc . wfs . v2_0_0 . catalog . source . WfsSourceTest . SAMPLE_FEATURE_NAME ) + "9" ) ) ; org . opengis . filter . Filter filter2 = builder . attribute ( Metacard . ANY_TEXT ) . is ( ) . like ( ) . text ( "*" ) ; org . opengis . filter . Filter typeSearchFilters = builder . anyOf ( filter0 , filter1 ) ; ddf . catalog . operation . impl . QueryImpl query = new ddf . catalog . operation . impl . QueryImpl ( builder . allOf ( filter2 , typeSearchFilters ) ) ; query . setPageSize ( pageSize ) ; net . opengis . wfs . v_2_0_0 . GetFeatureType featureType = source . buildGetFeatureRequest ( query ) ; int numTypes = featureType . getAbstractQueryExpression ( ) . size ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return provider . size ( ) ; } | org . junit . Assert . assertEquals ( 2 , numTypes ) |
checkAuthority_succeedsOnHostAndPort ( ) { java . lang . String actual = io . grpc . internal . GrpcUtil . checkAuthority ( "valid:1234" ) ; "<AssertPlaceHolder>" ; } checkAuthority ( java . lang . String ) { java . net . URI uri = io . grpc . internal . GrpcUtil . authorityToUri ( authority ) ; checkArgument ( ( ( uri . getHost ( ) ) != null ) , "No<sp>host<sp>in<sp>authority<sp>'%s'" , authority ) ; checkArgument ( ( ( uri . getUserInfo ( ) ) == null ) , "Userinfo<sp>must<sp>not<sp>be<sp>present<sp>on<sp>authority:<sp>'%s'" , authority ) ; return authority ; } | org . junit . Assert . assertEquals ( "valid:1234" , actual ) |
testSingleRow ( ) { hivemall . topicmodel . LDAUDTF udtf = new hivemall . topicmodel . LDAUDTF ( ) ; final int numTopics = 2 ; org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspector [ ] argOIs = new org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspector [ ] { org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspectorFactory . getStandardListObjectInspector ( PrimitiveObjectInspectorFactory . javaStringObjectInspector ) , org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspectorUtils . getConstantObjectInspector ( PrimitiveObjectInspectorFactory . javaStringObjectInspector , ( "-topics<sp>" + numTopics ) ) } ; udtf . initialize ( argOIs ) ; java . lang . String [ ] doc1 = new java . lang . String [ ] { "1" , "2" , "3" } ; udtf . process ( new java . lang . Object [ ] { java . util . Arrays . asList ( doc1 ) } ) ; final hivemall . utils . lang . mutable . MutableInt cnt = new hivemall . utils . lang . mutable . MutableInt ( 0 ) ; udtf . setCollector ( new org . apache . hadoop . hive . ql . udf . generic . Collector ( ) { @ hivemall . topicmodel . Override public void collect ( java . lang . Object arg0 ) throws org . apache . hadoop . hive . ql . metadata . HiveException { cnt . addValue ( 1 ) ; } } ) ; udtf . close ( ) ; "<AssertPlaceHolder>" ; } getValue ( ) { if ( ( covars ) == null ) { float w = weights [ cursor ] ; hivemall . model . WeightValue v = new hivemall . model . WeightValue ( w ) ; v . setTouched ( ( w != 0.0F ) ) ; return v ; } else { float w = weights [ cursor ] ; float cov = covars [ cursor ] ; hivemall . model . WeightValue . WeightValueWithCovar v = new hivemall . model . WeightValue . WeightValueWithCovar ( w , cov ) ; v . setTouched ( ( ( w != 0.0F ) || ( cov != 1.0F ) ) ) ; return v ; } } | org . junit . Assert . assertEquals ( ( ( doc1 . length ) * numTopics ) , cnt . getValue ( ) ) |
shouldReturnDefaultOffsetWhenOffsetIsNull ( ) { java . lang . Integer offset = org . openstack . atlas . api . helpers . PaginationHelper . determinePageOffset ( null ) ; "<AssertPlaceHolder>" ; } determinePageOffset ( java . lang . Integer ) { if ( offset == null ) return org . openstack . atlas . api . helpers . PaginationHelper . DEFAULT_PAGE_OFFSET ; else if ( offset < ( org . openstack . atlas . api . helpers . PaginationHelper . MIN_PAGE_OFFSET ) ) return org . openstack . atlas . api . helpers . PaginationHelper . MIN_PAGE_OFFSET ; else return offset ; } | org . junit . Assert . assertEquals ( PaginationHelper . DEFAULT_PAGE_OFFSET , offset ) |
testEmpty ( ) { java . util . Map < java . lang . Class < ? extends java . lang . annotation . Annotation > , java . util . List < com . google . gwt . core . ext . typeinfo . JClassType > > scanResult = com . mvp4g . rebind . AnnotationScanner . scan ( logger , oracle , com . mvp4g . rebind . TestAnnotationScanner . annotations ) ; for ( java . lang . Class < ? extends java . lang . annotation . Annotation > annotation : com . mvp4g . rebind . TestAnnotationScanner . annotations ) { "<AssertPlaceHolder>" ; } } scan ( com . google . gwt . core . ext . TreeLogger , com . google . gwt . core . ext . typeinfo . TypeOracle , com . mvp4g . rebind . Class [ ] ) { java . util . Map < java . lang . Class < ? extends java . lang . annotation . Annotation > , java . util . List < com . google . gwt . core . ext . typeinfo . JClassType > > annotationMap = new java . util . HashMap < java . lang . Class < ? extends java . lang . annotation . Annotation > , java . util . List < com . google . gwt . core . ext . typeinfo . JClassType > > ( ) ; for ( java . lang . Class < ? extends java . lang . annotation . Annotation > c : annotationClasses ) { annotationMap . put ( c , new java . util . ArrayList < com . google . gwt . core . ext . typeinfo . JClassType > ( ) ) ; } int nbClasses = 0 ; java . util . Date start = new java . util . Date ( ) ; for ( com . google . gwt . core . ext . typeinfo . JPackage pack : typeOracle . getPackages ( ) ) { for ( com . google . gwt . core . ext . typeinfo . JClassType type : pack . getTypes ( ) ) { for ( java . lang . Class < ? extends java . lang . annotation . Annotation > c : annotationClasses ) { if ( ( type . getAnnotation ( c ) ) != null ) { annotationMap . get ( c ) . add ( type ) ; } } } nbClasses += pack . getTypes ( ) . length ; } java . util . Date end = new java . util . Date ( ) ; logger . log ( TreeLogger . INFO , ( ( ( nbClasses + "<sp>classes<sp>scanned<sp>in<sp>" ) + ( java . lang . Long . toString ( ( ( end . getTime ( ) ) - ( start . getTime ( ) ) ) ) ) ) + "<sp>ms." ) ) ; return annotationMap ; } | org . junit . Assert . assertEquals ( 0 , scanResult . get ( annotation ) . size ( ) ) |
testExtendedValues ( ) { for ( int value : new int [ ] { 0 , 1 , 127 , 128 , 65536 , 198649 , Integer . MAX_VALUE } ) { java . nio . ByteBuffer buffer = encoder . encode ( value ) ; for ( int i = 1 ; i < 5 ; i ++ ) { int size = ( buffer . remaining ( ) ) + i ; org . apache . mina . codec . IoBuffer extendedBuffer = org . apache . mina . codec . IoBuffer . wrap ( java . nio . ByteBuffer . allocate ( size ) ) ; int start = extendedBuffer . position ( ) ; extendedBuffer . put ( buffer . slice ( ) ) ; extendedBuffer . position ( start ) ; extendedBuffer . limit ( ( start + size ) ) ; try { decoder . decode ( extendedBuffer ) ; "<AssertPlaceHolder>" ; } catch ( org . apache . mina . codec . ProtocolDecoderException e ) { org . junit . Assert . fail ( "Should<sp>not<sp>throw<sp>exception" ) ; } } } } remaining ( ) { return ( limit ( ) ) - ( position ( ) ) ; } | org . junit . Assert . assertEquals ( i , extendedBuffer . remaining ( ) ) |
testConvertToType07 ( ) { javax . el . CompositeELResolver resolver = new javax . el . CompositeELResolver ( ) ; resolver . add ( new javax . el . TesterELResolverOne ( ) ) ; resolver . add ( new javax . el . TesterELResolverTwo ( ) ) ; javax . el . ELContext context = new javax . el . TesterELContext ( resolver ) ; javax . el . ValueExpression ve = javax . el . ELManager . getExpressionFactory ( ) . createValueExpression ( context , "3" , java . lang . String . class ) ; java . lang . String result = ( ( java . lang . String ) ( ve . getValue ( context ) ) ) ; "<AssertPlaceHolder>" ; } getValue ( org . apache . el . lang . EvaluationContext ) { java . lang . Object obj = this . children [ 0 ] . getValue ( ctx ) ; if ( obj == null ) { return java . lang . Long . valueOf ( 0 ) ; } if ( obj instanceof java . math . BigDecimal ) { return ( ( java . math . BigDecimal ) ( obj ) ) . negate ( ) ; } if ( obj instanceof java . math . BigInteger ) { return ( ( java . math . BigInteger ) ( obj ) ) . negate ( ) ; } if ( obj instanceof java . lang . String ) { if ( isStringFloat ( ( ( java . lang . String ) ( obj ) ) ) ) { return new java . lang . Double ( ( - ( java . lang . Double . parseDouble ( ( ( java . lang . String ) ( obj ) ) ) ) ) ) ; } return java . lang . Long . valueOf ( ( - ( java . lang . Long . parseLong ( ( ( java . lang . String ) ( obj ) ) ) ) ) ) ; } if ( obj instanceof java . lang . Long ) { return java . lang . Long . valueOf ( ( - ( ( ( java . lang . Long ) ( obj ) ) . longValue ( ) ) ) ) ; } if ( obj instanceof java . lang . Double ) { return new java . lang . Double ( ( - ( ( ( java . lang . Double ) ( obj ) ) . doubleValue ( ) ) ) ) ; } if ( obj instanceof java . lang . Integer ) { return java . lang . Integer . valueOf ( ( - ( ( ( java . lang . Integer ) ( obj ) ) . intValue ( ) ) ) ) ; } if ( obj instanceof java . lang . Float ) { return new java . lang . Float ( ( - ( ( ( java . lang . Float ) ( obj ) ) . floatValue ( ) ) ) ) ; } if ( obj instanceof java . lang . Short ) { return java . lang . Short . valueOf ( ( ( short ) ( - ( ( ( java . lang . Short ) ( obj ) ) . shortValue ( ) ) ) ) ) ; } if ( obj instanceof java . lang . Byte ) { return java . lang . Byte . valueOf ( ( ( byte ) ( - ( ( ( java . lang . Byte ) ( obj ) ) . byteValue ( ) ) ) ) ) ; } java . lang . Long num = ( ( java . lang . Long ) ( coerceToNumber ( obj , org . apache . el . parser . Long . class ) ) ) ; return java . lang . Long . valueOf ( ( - ( num . longValue ( ) ) ) ) ; } | org . junit . Assert . assertEquals ( "3" , result ) |
testLong2Long ( ) { try { javax . jms . Message message = senderSession . createMessage ( ) ; message . setLongProperty ( "prop" , 127L ) ; "<AssertPlaceHolder>" ; } catch ( javax . jms . JMSException e ) { fail ( e ) ; } } getLongProperty ( java . lang . String ) { java . lang . Object prop = properties . get ( name ) ; if ( ! ( prop instanceof java . lang . Long ) ) { throw new javax . jms . JMSException ( "Not<sp>long" ) ; } return ( ( java . lang . Long ) ( properties . get ( name ) ) ) . longValue ( ) ; } | org . junit . Assert . assertEquals ( 127L , message . getLongProperty ( "prop" ) ) |
testCreateAndResizeDeltaNegative ( ) { java . lang . String filename = ( "/tmp/" + ( java . util . UUID . randomUUID ( ) ) ) + ".qcow2" ; long startSize = 81920 ; long increment = - 40960 ; org . apache . cloudstack . utils . qemu . QemuImgFile file = new org . apache . cloudstack . utils . qemu . QemuImgFile ( filename , startSize , org . apache . cloudstack . utils . qemu . QemuImg . PhysicalDiskFormat . RAW ) ; try { org . apache . cloudstack . utils . qemu . QemuImg qemu = new org . apache . cloudstack . utils . qemu . QemuImg ( 0 ) ; qemu . create ( file ) ; qemu . resize ( file , increment , true ) ; java . util . Map < java . lang . String , java . lang . String > info = qemu . info ( file ) ; if ( info == null ) { org . junit . Assert . fail ( "We<sp>didn't<sp>get<sp>any<sp>information<sp>back<sp>from<sp>qemu-img" ) ; } java . lang . Long infoSize = java . lang . Long . parseLong ( info . get ( new java . lang . String ( "virtual_size" ) ) ) ; "<AssertPlaceHolder>" ; } catch ( org . apache . cloudstack . utils . qemu . QemuImgException e ) { org . junit . Assert . fail ( e . getMessage ( ) ) ; } java . io . File f = new java . io . File ( filename ) ; f . delete ( ) ; } valueOf ( java . lang . String ) { java . lang . Number multiplier = 1 ; if ( ( multiplier ( ) ) != null ) { multiplier = ( ( java . lang . Number ) ( multiplier ( ) ) ) ; } java . lang . Class < T > type = type ( ) ; if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Boolean . class ) ) { return ( ( T ) ( java . lang . Boolean . valueOf ( value ) ) ) ; } else if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Integer . class ) ) { return ( ( T ) ( new java . lang . Integer ( ( ( java . lang . Integer . parseInt ( value ) ) * ( multiplier . intValue ( ) ) ) ) ) ) ; } else if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Long . class ) ) { return ( ( T ) ( new java . lang . Long ( ( ( java . lang . Long . parseLong ( value ) ) * ( multiplier . longValue ( ) ) ) ) ) ) ; } else if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Short . class ) ) { return ( ( T ) ( new java . lang . Short ( java . lang . Short . parseShort ( value ) ) ) ) ; } else if ( type . isAssignableFrom ( java . lang . String . class ) ) { return ( ( T ) ( value ) ) ; } else if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Float . class ) ) { return ( ( T ) ( new java . lang . Float ( ( ( java . lang . Float . parseFloat ( value ) ) * ( multiplier . floatValue ( ) ) ) ) ) ) ; } else if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Double . class ) ) { return ( ( T ) ( new java . lang . Double ( ( ( java . lang . Double . parseDouble ( value ) ) * ( multiplier . doubleValue ( ) ) ) ) ) ) ; } else if ( type . isAssignableFrom ( java . lang . String . class ) ) { return ( ( T ) ( value ) ) ; } else if ( type . isAssignableFrom ( java . sql . Date . class ) ) { return ( ( T ) ( java . sql . Date . valueOf ( value ) ) ) ; } else if ( type . isAssignableFrom ( org . apache . cloudstack . framework . config . Character . class ) ) { return ( ( T ) ( new java . lang . Character ( value . charAt ( 0 ) ) ) ) ; } else { throw new com . cloud . utils . exception . CloudRuntimeException ( ( "Unsupported<sp>data<sp>type<sp>for<sp>config<sp>values:<sp>" + type ) ) ; } } | org . junit . Assert . assertEquals ( java . lang . Long . valueOf ( ( startSize + increment ) ) , java . lang . Long . valueOf ( infoSize ) ) |
regexifyShouldGenerateSameValueForFakerWithSameSeed ( ) { long seed = 1L ; java . lang . String regex = "\\d" ; java . lang . String firstResult = new com . github . javafaker . Faker ( new java . util . Random ( seed ) ) . regexify ( regex ) ; java . lang . String secondResult = new com . github . javafaker . Faker ( new java . util . Random ( seed ) ) . regexify ( regex ) ; "<AssertPlaceHolder>" ; } regexify ( java . lang . String ) { return fakeValuesService . regexify ( regex ) ; } | org . junit . Assert . assertThat ( secondResult , is ( firstResult ) ) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.