diff --git a/src/tokenizer/tokenizer.rs b/src/tokenizer/tokenizer.rs index 84853fe32..b703c0b1b 100644 --- a/src/tokenizer/tokenizer.rs +++ b/src/tokenizer/tokenizer.rs @@ -88,7 +88,7 @@ impl TextAnalyzer { /// The resulting `BoxTokenStream` is equivalent to what would be obtained if the &str were /// one concatenated `&str`, with an artificial position gap of `2` between the different fields /// to prevent accidental `PhraseQuery` to match accross two terms. - pub fn token_stream_texts<'a>(&self, texts: &'a [&'a str]) -> Box { + pub fn token_stream_texts<'a>(&self, texts: &'a [&str]) -> Box { debug_assert!(!texts.is_empty()); let mut streams_with_offsets = vec![]; let mut total_offset = 0;