Skip to content

Commit e1d1d39

Browse files
Test Framework documentation (#80)
* Test Framework documentation * docs chatbot, split&embed, Prompts * prompt eng doc update * chatbot documentation page enhancement * split and embed documentation fixed * oci_config docs * final doc updates --------- Co-authored-by: John Lathouwers <[email protected]>
1 parent 867c385 commit e1d1d39

26 files changed

+231
-18
lines changed

docs/content/sandbox/chatbot/_index.md

Lines changed: 48 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,51 @@ Copyright (c) 2023, 2024, Oracle and/or its affiliates.
77
Licensed under the Universal Permissive License v1.0 as shown at http://oss.oracle.com/licenses/upl.
88
-->
99

10-
{{% notice style="code" title="10-Sept-2024: Documentation In-Progress..." icon="pen" %}}
11-
Thank you for your patience as we work on updating the documentation. Please check back soon for the latest updates.
12-
{{% /notice %}}
10+
Differently from a common LLM playground, that helps to test an LLM on the information on which has been trained on, the OAIM Sandbox works on the chunks retrieved in the Oracle DB 23ai by similarity with the question provided, like in this example:
11+
12+
![Chatbot](images/chatbot.png)
13+
14+
The playground could be used with or without the vector stores available, to check if a pure LLM configured is aware or not about information you are looking for.
15+
16+
You can, first of all:
17+
18+
- **Enable History and Context**: in this way any further question & answer provided will be re-sent in the context to help the LLM to answer with a better grounded info, if it's checked;
19+
- **Clear History**: this button clear the context to better understand the LLM behaviour after a long conversation;
20+
21+
## Chat Model
22+
Depending on the configuration done in the **Configuration**/**Models** page, you can choose one of the **Chat model** enlisted. For each of them you can modify the most important hyper-parameters like:
23+
- Temperature
24+
- Maximum Tokens
25+
- Top P
26+
- Frequency penalty
27+
- Presence penalty
28+
29+
To understand each of them, refers for example on this document: [Concepts for Generative AI](https://docs.oracle.com/en-us/iaas/Content/generative-ai/concepts.htm).
30+
31+
## RAG params
32+
33+
Clicking on the **RAG** checkbox you can quickly turn on/off the knowledge base behind the chatbot, exploiting the Retrieval Augentened Generation pattern implemented into the Oracle AI Microserves Sandbox.
34+
35+
![Playground](images/playground.png)
36+
37+
Then you can set:
38+
39+
- **Enable Re-Ranking**: *under development*;
40+
- **Search Type**: it reflects the two options available on the **Oracle DB 23ai**:
41+
- **Similarity search**
42+
- **Maximal Marginal Relevance**.
43+
- **Top K**: define the number of nearest chunks, found comparing the embedding vector derived by the question with the vectors associated in the vectorstore with each chunk. Take in consideration that a large number of chunk could fill the maximum context size accepted by the LLM becoming useless the text that exceeds that limit.
44+
45+
To search and select one of the vectorstore tables created into the DB and use it for the RAG, you could use one, or the combination of more than one parameter adopted in the chunking process, to filter the desired vectorstore:
46+
47+
- **Embedding Alias**
48+
- **Embedding Model**
49+
- **Chunk Size**
50+
- **Chunk Overlap**
51+
- **Distance Strategy**
52+
53+
Until the following message will not disappear, it means that the final vectorstore is not yet selected:
54+
55+
![Rag Error message](images/ragmessage.png)
56+
57+
The **Reset RAG** button allows you to restart the selection of another vectorestore table.
56.6 KB
Loading
35 KB
Loading
43.4 KB
Loading
55.1 KB
Loading
45.7 KB
Loading
71.3 KB
Loading
14 KB
Loading
53.2 KB
Loading
6.7 KB
Loading

0 commit comments

Comments
 (0)