react-code-dataset
/
wp-calypso
/client
/state
/data-layer
/wpcom-http
/pipeline
/remove-duplicate-gets
/test
/index.js
| import { | |
| addResponder, | |
| applyDuplicatesHandlers, | |
| buildKey, | |
| clearQueue, | |
| removeDuplicateGets, | |
| } from '../'; | |
| const failer = { type: 'FAIL' }; | |
| const filler = { type: 'FILL' }; | |
| const succeeder = { type: 'SUCCEED' }; | |
| const getSites = { | |
| method: 'GET', | |
| path: '/sites', | |
| apiVersion: 'v1', | |
| onSuccess: succeeder, | |
| onFailure: failer, | |
| }; | |
| const getPosts = { | |
| method: 'GET', | |
| path: '/sites/posts', | |
| apiVersion: 'v1.2', | |
| onSuccess: succeeder, | |
| onFailure: failer, | |
| }; | |
| const postLike = { | |
| method: 'POST', | |
| path: '/sites/posts/like', | |
| apiVersion: 'v1.4', | |
| onSuccess: succeeder, | |
| onFailure: failer, | |
| }; | |
| /** | |
| * Make a quick clone of an object for testing | |
| * @param {Object} o object to shallow-clone | |
| * @returns {Object} cloned object | |
| */ | |
| const cp = ( o ) => ( { ...o } ); | |
| describe( '#buildKey', () => { | |
| test( 'should collapse "duplicate" requests', () => { | |
| const duplicates = [ | |
| [ | |
| { path: '/', apiNamespace: 'wpcom/v2', query: { id: 1, limit: 5 } }, | |
| { path: '/', apiNamespace: 'wpcom/v2', query: { limit: 5, id: 1 } }, | |
| ], | |
| [ { path: '/' }, { path: '/', query: undefined } ], | |
| [ { path: '/' }, { path: '/', query: {} } ], | |
| [ { path: '/' }, { path: '/', apiNamespace: undefined } ], | |
| [ | |
| { path: '/', onSuccess: succeeder }, | |
| { path: '/', onSuccess: filler }, | |
| ], | |
| ]; | |
| duplicates.forEach( ( [ a, b ] ) => expect( buildKey( a ) ).toEqual( buildKey( b ) ) ); | |
| } ); | |
| test( 'should differentiate "unique" requests', () => { | |
| const uniques = [ | |
| [ | |
| { path: '/', apiNamespace: 'wp/v1' }, | |
| { path: '/', apiNamespace: 'wpcom/v1' }, | |
| ], | |
| [ | |
| { path: '/', apiNamespace: 'wp/v1' }, | |
| { path: '/', apiVersion: 'wp/v1' }, | |
| ], | |
| [ { path: '/' }, { path: '/a' } ], | |
| [ | |
| { path: '/', query: { id: 1 } }, | |
| { path: '/', query: { id: 2 } }, | |
| ], | |
| ]; | |
| uniques.forEach( ( [ a, b ] ) => expect( buildKey( a ) ).not.toEqual( buildKey( b ) ) ); | |
| } ); | |
| } ); | |
| describe( '#addResponder', () => { | |
| test( 'should add an `onFailure` action to an empty list', () => { | |
| const union = addResponder( {}, { onFailure: cp( failer ) } ); | |
| expect( union.failures ).toEqual( [ failer ] ); | |
| expect( union.successes ).toHaveLength( 0 ); | |
| } ); | |
| test( 'should add an `onSuccess` action to an empty list', () => { | |
| const union = addResponder( {}, { onSuccess: cp( succeeder ) } ); | |
| expect( union.failures ).toHaveLength( 0 ); | |
| expect( union.successes ).toEqual( [ succeeder ] ); | |
| } ); | |
| test( 'should add a "unique" action to an existing list', () => { | |
| const union = addResponder( { successes: [ cp( succeeder ) ] }, { onSuccess: cp( filler ) } ); | |
| expect( union.successes ).toEqual( [ succeeder, filler ] ); | |
| } ); | |
| test( 'should merge "duplicate" actions to an existing list', () => { | |
| const union = addResponder( | |
| { successes: [ cp( succeeder ) ] }, | |
| { onSuccess: cp( succeeder ) } | |
| ); | |
| expect( union.successes ).toEqual( [ succeeder ] ); | |
| } ); | |
| test( 'should add both `onSuccess` and `onFailure`', () => { | |
| const union = addResponder( | |
| { | |
| failures: [ cp( failer ) ], | |
| successes: [ cp( succeeder ) ], | |
| }, | |
| { | |
| onFailure: cp( filler ), | |
| onSuccess: cp( filler ), | |
| } | |
| ); | |
| expect( union.failures ).toEqual( [ failer, filler ] ); | |
| expect( union.successes ).toEqual( [ succeeder, filler ] ); | |
| } ); | |
| } ); | |
| describe( '#removeDuplicateGets', () => { | |
| beforeEach( clearQueue ); | |
| test( 'should pass through non-GET requests', () => { | |
| const primed = removeDuplicateGets( { nextRequest: cp( postLike ) } ); | |
| expect( primed.nextRequest ).toEqual( postLike ); | |
| const processed = removeDuplicateGets( { nextRequest: cp( postLike ) } ); | |
| expect( processed.nextRequest ).toEqual( postLike ); | |
| } ); | |
| test( 'should pass through new requests', () => { | |
| const processed = removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| expect( processed.nextRequest ).toEqual( getSites ); | |
| } ); | |
| test( 'should pass through "unique" requests', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const processed = removeDuplicateGets( { nextRequest: cp( getPosts ) } ); | |
| expect( processed.nextRequest ).toEqual( getPosts ); | |
| } ); | |
| test( 'should drop "duplicate" requests', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const processed = removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| expect( processed.nextRequest ).toBeNull(); | |
| } ); | |
| } ); | |
| describe( '#applyDuplicateHandlers', () => { | |
| beforeAll( clearQueue ); | |
| test( 'should return new requests', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const processed = applyDuplicatesHandlers( { originalRequest: cp( getSites ) } ); | |
| expect( processed.failures ).toEqual( [ getSites.onFailure ] ); | |
| expect( processed.successes ).toEqual( [ getSites.onSuccess ] ); | |
| } ); | |
| test( 'should collapse "duplicate" requests having same responders', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const processed = applyDuplicatesHandlers( { originalRequest: cp( getSites ) } ); | |
| expect( processed.failures ).toEqual( [ getSites.onFailure ] ); | |
| expect( processed.successes ).toEqual( [ getSites.onSuccess ] ); | |
| } ); | |
| test( 'should spread "duplicate" requests having different responders', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| removeDuplicateGets( { nextRequest: { ...getSites, onSuccess: cp( filler ) } } ); | |
| const processed = applyDuplicatesHandlers( { originalRequest: cp( getSites ) } ); | |
| expect( processed.failures ).toEqual( [ getSites.onFailure ] ); | |
| expect( processed.successes ).toEqual( [ getSites.onSuccess, filler ] ); | |
| } ); | |
| test( 'should pass through "unique" requests', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| removeDuplicateGets( { nextRequest: cp( getPosts ) } ); | |
| const sites = applyDuplicatesHandlers( { originalRequest: cp( getSites ) } ); | |
| expect( sites.failures ).toEqual( [ failer ] ); | |
| expect( sites.successes ).toEqual( [ succeeder ] ); | |
| const posts = applyDuplicatesHandlers( { originalRequest: cp( getPosts ) } ); | |
| expect( posts.failures ).toEqual( [ failer ] ); | |
| expect( posts.successes ).toEqual( [ succeeder ] ); | |
| } ); | |
| test( 'should pass through "duplicate" requests which never overlap', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const first = applyDuplicatesHandlers( { originalRequest: cp( getSites ) } ); | |
| expect( first.failures ).toEqual( [ failer ] ); | |
| expect( first.successes ).toEqual( [ succeeder ] ); | |
| const { nextRequest } = removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| expect( nextRequest ).toEqual( getSites ); | |
| const second = applyDuplicatesHandlers( { originalRequest: cp( getSites ) } ); | |
| expect( second.failures ).toEqual( [ failer ] ); | |
| expect( second.successes ).toEqual( [ succeeder ] ); | |
| } ); | |
| test( 'should not collapse non-GET requests', () => { | |
| removeDuplicateGets( { nextRequest: cp( postLike ) } ); | |
| removeDuplicateGets( { nextRequest: cp( postLike ) } ); | |
| const processed = applyDuplicatesHandlers( { | |
| failures: [ cp( postLike.onFailure ) ], | |
| originalRequest: cp( postLike ), | |
| successes: [ cp( postLike.onSuccess ) ], | |
| } ); | |
| expect( processed.failures ).toEqual( [ postLike.onFailure ] ); | |
| expect( processed.successes ).toEqual( [ postLike.onSuccess ] ); | |
| } ); | |
| test( 'should not wipe out previous responders in the pipeline', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const processed = applyDuplicatesHandlers( { | |
| failures: [ cp( filler ) ], | |
| originalRequest: cp( getSites ), | |
| successes: [ cp( filler ) ], | |
| } ); | |
| expect( processed.failures ).toEqual( [ filler, getSites.onFailure ] ); | |
| expect( processed.successes ).toEqual( [ filler, getSites.onSuccess ] ); | |
| } ); | |
| test( 'should combine previous responders with "duplicate" responders in the pipeline', () => { | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| removeDuplicateGets( { nextRequest: cp( getSites ) } ); | |
| const processed = applyDuplicatesHandlers( { | |
| failures: [ cp( getSites.onFailure ) ], | |
| originalRequest: cp( getSites ), | |
| successes: [ cp( getSites.onSuccess ) ], | |
| } ); | |
| expect( processed.failures ).toEqual( [ getSites.onFailure ] ); | |
| expect( processed.successes ).toEqual( [ getSites.onSuccess ] ); | |
| } ); | |
| } ); | |