|
|
|
@ -57,7 +57,7 @@ public class Jackson2TokenizerTests extends AbstractDataBufferAllocatingTestCase |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
@Test |
|
|
|
@Test |
|
|
|
public void noTokenizeArrayElements() { |
|
|
|
public void doNotTokenizeArrayElements() { |
|
|
|
this.tokenizer = new Jackson2Tokenizer(this.jsonParser, false); |
|
|
|
this.tokenizer = new Jackson2Tokenizer(this.jsonParser, false); |
|
|
|
|
|
|
|
|
|
|
|
testTokenize( |
|
|
|
testTokenize( |
|
|
|
@ -116,7 +116,7 @@ public class Jackson2TokenizerTests extends AbstractDataBufferAllocatingTestCase |
|
|
|
asList("{\"foo\": \"bar\"}", |
|
|
|
asList("{\"foo\": \"bar\"}", |
|
|
|
"{\"foo\": \"baz\"}")); |
|
|
|
"{\"foo\": \"baz\"}")); |
|
|
|
|
|
|
|
|
|
|
|
// SPR-15803
|
|
|
|
// SPR-15803: nested array
|
|
|
|
testTokenize( |
|
|
|
testTokenize( |
|
|
|
singletonList("[" + |
|
|
|
singletonList("[" + |
|
|
|
"{\"id\":\"0\",\"start\":[-999999999,1,1],\"end\":[999999999,12,31]}," + |
|
|
|
"{\"id\":\"0\",\"start\":[-999999999,1,1],\"end\":[999999999,12,31]}," + |
|
|
|
@ -129,6 +129,11 @@ public class Jackson2TokenizerTests extends AbstractDataBufferAllocatingTestCase |
|
|
|
"{\"id\":\"2\",\"start\":[-999999999,1,1],\"end\":[999999999,12,31]}") |
|
|
|
"{\"id\":\"2\",\"start\":[-999999999,1,1],\"end\":[999999999,12,31]}") |
|
|
|
); |
|
|
|
); |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// SPR-15803: nested array, no top-level array
|
|
|
|
|
|
|
|
testTokenize( |
|
|
|
|
|
|
|
singletonList("{\"speakerIds\":[\"tastapod\"],\"language\":\"ENGLISH\"}"), |
|
|
|
|
|
|
|
singletonList("{\"speakerIds\":[\"tastapod\"],\"language\":\"ENGLISH\"}")); |
|
|
|
|
|
|
|
|
|
|
|
testTokenize( |
|
|
|
testTokenize( |
|
|
|
asList("[{\"foo\": \"foofoo\", \"bar\"", |
|
|
|
asList("[{\"foo\": \"foofoo\", \"bar\"", |
|
|
|
": \"barbar\"},{\"foo\": \"foofoofoo\", \"bar\": \"barbarbar\"}]"), |
|
|
|
": \"barbar\"},{\"foo\": \"foofoofoo\", \"bar\": \"barbarbar\"}]"), |
|
|
|
|