-
Notifications
You must be signed in to change notification settings - Fork 316
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
feat(fulltext_index): introduce creator (#4249)
* feat(fulltext_index): introduce creator Signed-off-by: Zhenchi <[email protected]> * fix: typo Signed-off-by: Zhenchi <[email protected]> * fix: typo Signed-off-by: Zhenchi <[email protected]> * chore: polish Signed-off-by: Zhenchi <[email protected]> * fix: return error if writer not found Signed-off-by: Zhenchi <[email protected]> * refactor: helper function for tests Signed-off-by: Zhenchi <[email protected]> --------- Signed-off-by: Zhenchi <[email protected]>
- Loading branch information
Showing
9 changed files
with
735 additions
and
16 deletions.
There are no files selected for viewing
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,37 @@ | ||
// Copyright 2023 Greptime Team | ||
// | ||
// Licensed under the Apache License, Version 2.0 (the "License"); | ||
// you may not use this file except in compliance with the License. | ||
// You may obtain a copy of the License at | ||
// | ||
// http://www.apache.org/licenses/LICENSE-2.0 | ||
// | ||
// Unless required by applicable law or agreed to in writing, software | ||
// distributed under the License is distributed on an "AS IS" BASIS, | ||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
// See the License for the specific language governing permissions and | ||
// limitations under the License. | ||
|
||
use serde::{Deserialize, Serialize}; | ||
|
||
pub mod create; | ||
pub mod error; | ||
|
||
/// Configuration for fulltext index. | ||
#[derive(Debug, Clone, PartialEq, Eq, Serialize, Deserialize, Default)] | ||
pub struct Config { | ||
/// Analyzer to use for tokenization. | ||
pub analyzer: Analyzer, | ||
|
||
/// Whether the index should be case-sensitive. | ||
pub case_sensitive: bool, | ||
} | ||
|
||
/// Analyzer to use for tokenization. | ||
#[derive(Debug, Clone, PartialEq, Eq, Serialize, Deserialize, Default)] | ||
pub enum Analyzer { | ||
#[default] | ||
English, | ||
|
||
Chinese, | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,33 @@ | ||
// Copyright 2023 Greptime Team | ||
// | ||
// Licensed under the Apache License, Version 2.0 (the "License"); | ||
// you may not use this file except in compliance with the License. | ||
// You may obtain a copy of the License at | ||
// | ||
// http://www.apache.org/licenses/LICENSE-2.0 | ||
// | ||
// Unless required by applicable law or agreed to in writing, software | ||
// distributed under the License is distributed on an "AS IS" BASIS, | ||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
// See the License for the specific language governing permissions and | ||
// limitations under the License. | ||
|
||
mod tantivy; | ||
|
||
use async_trait::async_trait; | ||
pub use tantivy::TantivyFulltextIndexCreator; | ||
|
||
use crate::fulltext_index::error::Result; | ||
|
||
/// `FulltextIndexCreator` is for creating a fulltext index. | ||
#[async_trait] | ||
pub trait FulltextIndexCreator: Send { | ||
/// Pushes a text to the index. | ||
async fn push_text(&mut self, text: &str) -> Result<()>; | ||
|
||
/// Finalizes the creation of the index. | ||
async fn finish(&mut self) -> Result<()>; | ||
|
||
/// Returns the memory usage in bytes during the creation of the index. | ||
fn memory_usage(&self) -> usize; | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,243 @@ | ||
// Copyright 2023 Greptime Team | ||
// | ||
// Licensed under the Apache License, Version 2.0 (the "License"); | ||
// you may not use this file except in compliance with the License. | ||
// You may obtain a copy of the License at | ||
// | ||
// http://www.apache.org/licenses/LICENSE-2.0 | ||
// | ||
// Unless required by applicable law or agreed to in writing, software | ||
// distributed under the License is distributed on an "AS IS" BASIS, | ||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
// See the License for the specific language governing permissions and | ||
// limitations under the License. | ||
|
||
use std::path::Path; | ||
|
||
use async_trait::async_trait; | ||
use snafu::{OptionExt, ResultExt}; | ||
use tantivy::schema::{Schema, TEXT}; | ||
use tantivy::store::{Compressor, ZstdCompressor}; | ||
use tantivy::tokenizer::{LowerCaser, SimpleTokenizer, TextAnalyzer, TokenizerManager}; | ||
use tantivy::{doc, Index, SingleSegmentIndexWriter}; | ||
use tantivy_jieba::JiebaTokenizer; | ||
|
||
use crate::fulltext_index::create::FulltextIndexCreator; | ||
use crate::fulltext_index::error::{FinishedSnafu, IoSnafu, Result, TantivySnafu}; | ||
use crate::fulltext_index::{Analyzer, Config}; | ||
|
||
pub const TEXT_FIELD_NAME: &str = "greptime_fulltext_text"; | ||
|
||
/// `TantivyFulltextIndexCreator` is a fulltext index creator using tantivy. | ||
/// | ||
/// Here use a single segment to store the index so the maximum capacity for | ||
/// the index is limited to 2<<31 rows (around 2 billion rows). | ||
pub struct TantivyFulltextIndexCreator { | ||
/// The tantivy index writer. | ||
writer: Option<SingleSegmentIndexWriter>, | ||
|
||
/// The field for the text. | ||
text_field: tantivy::schema::Field, | ||
} | ||
|
||
impl TantivyFulltextIndexCreator { | ||
/// Creates a new `TantivyFulltextIndexCreator`. | ||
/// | ||
/// The `path` is the directory path in filesystem to store the index. | ||
pub async fn new(path: impl AsRef<Path>, config: Config, memory_limit: usize) -> Result<Self> { | ||
tokio::fs::create_dir_all(path.as_ref()) | ||
.await | ||
.context(IoSnafu)?; | ||
|
||
let mut schema_builder = Schema::builder(); | ||
let text_field = schema_builder.add_text_field(TEXT_FIELD_NAME, TEXT); | ||
let schema = schema_builder.build(); | ||
|
||
let mut index = Index::create_in_dir(path, schema).context(TantivySnafu)?; | ||
index.settings_mut().docstore_compression = Compressor::Zstd(ZstdCompressor::default()); | ||
index.set_tokenizers(Self::build_tokenizer(&config)); | ||
|
||
let memory_limit = Self::sanitize_memory_limit(memory_limit); | ||
|
||
let writer = SingleSegmentIndexWriter::new(index, memory_limit).context(TantivySnafu)?; | ||
Ok(Self { | ||
writer: Some(writer), | ||
text_field, | ||
}) | ||
} | ||
|
||
fn build_tokenizer(config: &Config) -> TokenizerManager { | ||
let mut builder = match config.analyzer { | ||
Analyzer::English => TextAnalyzer::builder(SimpleTokenizer::default()).dynamic(), | ||
Analyzer::Chinese => TextAnalyzer::builder(JiebaTokenizer {}).dynamic(), | ||
}; | ||
|
||
if !config.case_sensitive { | ||
builder = builder.filter_dynamic(LowerCaser); | ||
} | ||
|
||
let tokenizer = builder.build(); | ||
let tokenizer_manager = TokenizerManager::new(); | ||
tokenizer_manager.register("default", tokenizer); | ||
tokenizer_manager | ||
} | ||
|
||
fn sanitize_memory_limit(memory_limit: usize) -> usize { | ||
// Port from tantivy::indexer::index_writer::{MEMORY_BUDGET_NUM_BYTES_MIN, MEMORY_BUDGET_NUM_BYTES_MAX} | ||
const MARGIN_IN_BYTES: usize = 1_000_000; | ||
const MEMORY_BUDGET_NUM_BYTES_MIN: usize = ((MARGIN_IN_BYTES as u32) * 15u32) as usize; | ||
const MEMORY_BUDGET_NUM_BYTES_MAX: usize = u32::MAX as usize - MARGIN_IN_BYTES; | ||
|
||
memory_limit.clamp(MEMORY_BUDGET_NUM_BYTES_MIN, MEMORY_BUDGET_NUM_BYTES_MAX) | ||
} | ||
} | ||
|
||
#[async_trait] | ||
impl FulltextIndexCreator for TantivyFulltextIndexCreator { | ||
async fn push_text(&mut self, text: &str) -> Result<()> { | ||
let writer = self.writer.as_mut().context(FinishedSnafu)?; | ||
let doc = doc!(self.text_field => text); | ||
writer.add_document(doc).context(TantivySnafu) | ||
} | ||
|
||
async fn finish(&mut self) -> Result<()> { | ||
let writer = self.writer.take().context(FinishedSnafu)?; | ||
writer.finalize().map(|_| ()).context(TantivySnafu) | ||
} | ||
|
||
fn memory_usage(&self) -> usize { | ||
self.writer | ||
.as_ref() | ||
.map(|writer| writer.mem_usage()) | ||
.unwrap_or(0) | ||
} | ||
} | ||
|
||
#[cfg(test)] | ||
mod tests { | ||
use common_test_util::temp_dir::create_temp_dir; | ||
use tantivy::collector::DocSetCollector; | ||
use tantivy::query::QueryParser; | ||
|
||
use super::*; | ||
|
||
#[tokio::test] | ||
async fn test_creator_basic() { | ||
let memory_limits = [1, 64_000_000, usize::MAX]; | ||
|
||
for memory_limit in memory_limits { | ||
let temp_dir = create_temp_dir("test_creator_basic_"); | ||
|
||
let texts = vec!["hello", "world", "hello, world", "foo!", "Bar"]; | ||
let config = Config::default(); | ||
build_index(&texts, temp_dir.path(), config, memory_limit).await; | ||
|
||
let cases = [ | ||
("hello", vec![0u32, 2]), | ||
("world", vec![1, 2]), | ||
("foo", vec![3]), | ||
("bar", vec![4]), | ||
]; | ||
query_and_check(temp_dir.path(), &cases).await; | ||
} | ||
} | ||
|
||
#[tokio::test] | ||
async fn test_creator_case_sensitive() { | ||
let memory_limits = [1, 64_000_000, usize::MAX]; | ||
|
||
for memory_limit in memory_limits { | ||
let temp_dir = create_temp_dir("test_creator_case_sensitive_"); | ||
|
||
let texts = vec!["hello", "world", "hello, world", "foo!", "Bar"]; | ||
let config = Config { | ||
case_sensitive: true, | ||
..Config::default() | ||
}; | ||
build_index(&texts, temp_dir.path(), config, memory_limit).await; | ||
|
||
let cases = [ | ||
("hello", vec![0u32, 2]), | ||
("world", vec![1, 2]), | ||
("foo", vec![3]), | ||
("bar", vec![]), | ||
]; | ||
query_and_check(temp_dir.path(), &cases).await; | ||
} | ||
} | ||
|
||
#[tokio::test] | ||
async fn test_creator_chinese() { | ||
let memory_limits = [1, 64_000_000, usize::MAX]; | ||
|
||
for memory_limit in memory_limits { | ||
let temp_dir = create_temp_dir("test_creator_chinese_"); | ||
|
||
let texts = vec!["你好", "世界", "你好,世界", "你好世界", "foo!", "Bar"]; | ||
let config = Config { | ||
analyzer: Analyzer::Chinese, | ||
..Config::default() | ||
}; | ||
build_index(&texts, temp_dir.path(), config, memory_limit).await; | ||
|
||
let cases = [ | ||
("你好", vec![0u32, 2, 3]), | ||
("世界", vec![1, 2, 3]), | ||
("foo", vec![4]), | ||
("bar", vec![5]), | ||
]; | ||
query_and_check(temp_dir.path(), &cases).await; | ||
} | ||
} | ||
|
||
#[tokio::test] | ||
async fn test_creator_chinese_case_sensitive() { | ||
let memory_limits = [1, 64_000_000, usize::MAX]; | ||
|
||
for memory_limit in memory_limits { | ||
let temp_dir = create_temp_dir("test_creator_chinese_case_sensitive_"); | ||
|
||
let texts = vec!["你好", "世界", "你好,世界", "你好世界", "foo!", "Bar"]; | ||
let config = Config { | ||
case_sensitive: true, | ||
analyzer: Analyzer::Chinese, | ||
}; | ||
build_index(&texts, temp_dir.path(), config, memory_limit).await; | ||
|
||
let cases = [ | ||
("你好", vec![0u32, 2, 3]), | ||
("世界", vec![1, 2, 3]), | ||
("foo", vec![4]), | ||
("bar", vec![]), | ||
]; | ||
query_and_check(temp_dir.path(), &cases).await; | ||
} | ||
} | ||
|
||
async fn build_index(texts: &[&str], path: &Path, config: Config, memory_limit: usize) { | ||
let mut creator = TantivyFulltextIndexCreator::new(path, config, memory_limit) | ||
.await | ||
.unwrap(); | ||
for text in texts { | ||
creator.push_text(text).await.unwrap(); | ||
} | ||
creator.finish().await.unwrap(); | ||
} | ||
|
||
async fn query_and_check(path: &Path, cases: &[(&str, Vec<u32>)]) { | ||
let index = Index::open_in_dir(path).unwrap(); | ||
let reader = index.reader().unwrap(); | ||
let searcher = reader.searcher(); | ||
for (query, expected) in cases { | ||
let query_parser = QueryParser::for_index( | ||
&index, | ||
vec![index.schema().get_field(TEXT_FIELD_NAME).unwrap()], | ||
); | ||
let query = query_parser.parse_query(query).unwrap(); | ||
let docs = searcher.search(&query, &DocSetCollector).unwrap(); | ||
let mut res = docs.into_iter().map(|d| d.doc_id).collect::<Vec<_>>(); | ||
res.sort(); | ||
assert_eq!(res, *expected); | ||
} | ||
} | ||
} |
Oops, something went wrong.