diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md deleted file mode 100644 index 099adaa04..000000000 --- a/CONTRIBUTING.md +++ /dev/null @@ -1,5 +0,0 @@ -# Contributing to TruLens Eval -[Contribution Guidelines](https://github.com/truera/trulens/blob/main/trulens_eval/CONTRIBUTING.md) - -# Contributing to TruLens Explain -[Contribution Guidelines](https://github.com/truera/trulens/blob/main/trulens_explain/CONTRIBUTING.md) diff --git a/README.md b/README.md index 8a1739674..bc237078a 100644 --- a/README.md +++ b/README.md @@ -20,7 +20,7 @@ The best way to support TruLens is to give us a ⭐ on [GitHub](https://www.gith Fine-grained, stack-agnostic instrumentation and comprehensive evaluations help you to identify failure modes & systematically iterate to improve your application. -Read more about the core concepts behind TruLens including [Feedback Functions](./trulens_eval/core_concepts_feedback_functions.md), [The RAG Triad](./core_concepts_rag_triad.md), and [Honest, Harmless and Helpful Evals](./core_concepts_honest_harmless_helpful_evals.md). +Read more about the core concepts behind TruLens including [Feedback Functions](https://www.trulens.org/trulens_eval/core_concepts_feedback_functions/), [The RAG Triad](https://www.trulens.org/trulens_eval/core_concepts_rag_triad/), and [Honest, Harmless and Helpful Evals](https://www.trulens.org/trulens_eval/core_concepts_honest_harmless_helpful_evals/). ## TruLens in the development workflow @@ -44,7 +44,7 @@ Walk through how to instrument and evaluate a RAG built from scratch with TruLen ### 💡 Contributing -Interested in contributing? See our [contribution guide](https://github.com/truera/trulens/tree/main/trulens_eval/CONTRIBUTING.md) for more details. +Interested in contributing? See our [contribution guide](https://www.trulens.org/trulens_eval/CONTRIBUTING/) for more details. ## TruLens-Explain diff --git a/docs/trulens_eval/core_concepts_feedback_functions.md b/docs/trulens_eval/core_concepts_feedback_functions.md index 605d4db93..eb4163461 100644 --- a/docs/trulens_eval/core_concepts_feedback_functions.md +++ b/docs/trulens_eval/core_concepts_feedback_functions.md @@ -10,13 +10,13 @@ It can be useful to think of the range of evaluations on two axis: Scalable and In early development stages, we recommend starting with domain expert evaluations. These evaluations are often completed by the developers themselves and represent the core use cases your app is expected to complete. This allows you to deeply understand the performance of your app, but lacks scale. -See this [example notebook](./groundtruth_evals.ipynb) to learn how to run ground truth evaluations with TruLens. +See this [example notebook](https://www.trulens.org/trulens_eval/groundtruth_evals/) to learn how to run ground truth evaluations with TruLens. ## User Feedback (Human) Evaluations After you have completed early evaluations and have gained more confidence in your app, it is often useful to gather human feedback. This can often be in the form of binary (up/down) feedback provided by your users. This is more slightly scalable than ground truth evals, but struggles with variance and can still be expensive to collect. -See this [example notebook](./human_feedback.ipynb) to learn how to log human feedback with TruLens. +See this [example notebook](https://www.trulens.org/trulens_eval/human_feedback/) to learn how to log human feedback with TruLens. ## Traditional NLP Evaluations @@ -34,4 +34,4 @@ Large Language Models can also provide meaningful and flexible feedback on LLM a Depending on the size and nature of the LLM, these evaluations can be quite expensive at scale. -See this [example notebook](./quickstart.ipynb) to learn how to run LLM-based evaluations with TruLens. \ No newline at end of file +See this [example notebook](https://www.trulens.org/trulens_eval/quickstart/) to learn how to run LLM-based evaluations with TruLens. \ No newline at end of file diff --git a/docs/trulens_eval/core_concepts_rag_triad.md b/docs/trulens_eval/core_concepts_rag_triad.md index 9b4703a6f..3c1035fbc 100644 --- a/docs/trulens_eval/core_concepts_rag_triad.md +++ b/docs/trulens_eval/core_concepts_rag_triad.md @@ -24,5 +24,5 @@ Last, our response still needs to helpfully answer the original question. We can By reaching satisfactory evaluations for this triad, we can make a nuanced statement about our application’s correctness; our application is verified to be hallucination free up to the limit of its knowledge base. In other words, if the vector database contains only accurate information, then the answers provided by the RAG are also accurate. -To see the RAG triad in action, check out the [TruLens Quickstart](./quickstart.ipynb) +To see the RAG triad in action, check out the [TruLens Quickstart](https://www.trulens.org/trulens_eval/quickstart/) diff --git a/docs/trulens_eval/gh_top_intro.md b/docs/trulens_eval/gh_top_intro.md index 56f2e36d8..cd9da0d69 100644 --- a/docs/trulens_eval/gh_top_intro.md +++ b/docs/trulens_eval/gh_top_intro.md @@ -20,7 +20,7 @@ The best way to support TruLens is to give us a ⭐ on [GitHub](https://www.gith Fine-grained, stack-agnostic instrumentation and comprehensive evaluations help you to identify failure modes & systematically iterate to improve your application. -Read more about the core concepts behind TruLens including [Feedback Functions](./trulens_eval/core_concepts_feedback_functions.md), [The RAG Triad](./core_concepts_rag_triad.md), and [Honest, Harmless and Helpful Evals](./core_concepts_honest_harmless_helpful_evals.md). +Read more about the core concepts behind TruLens including [Feedback Functions](https://www.trulens.org/trulens_eval/core_concepts_feedback_functions/), [The RAG Triad](https://www.trulens.org/trulens_eval/core_concepts_rag_triad/), and [Honest, Harmless and Helpful Evals](https://www.trulens.org/trulens_eval/core_concepts_honest_harmless_helpful_evals/). ## TruLens in the development workflow @@ -44,4 +44,4 @@ Walk through how to instrument and evaluate a RAG built from scratch with TruLen ### 💡 Contributing -Interested in contributing? See our [contribution guide](https://github.com/truera/trulens/tree/main/trulens_eval/CONTRIBUTING.md) for more details. +Interested in contributing? See our [contribution guide](https://www.trulens.org/trulens_eval/CONTRIBUTING/) for more details. diff --git a/docs/trulens_eval/intro.md b/docs/trulens_eval/intro.md index 078f95ed8..fa4c99997 100644 --- a/docs/trulens_eval/intro.md +++ b/docs/trulens_eval/intro.md @@ -6,7 +6,7 @@ Fine-grained, stack-agnostic instrumentation and comprehensive evaluations help you to identify failure modes & systematically iterate to improve your application. -Read more about the core concepts behind TruLens including [Feedback Functions](./trulens_eval/core_concepts_feedback_functions.md), [The RAG Triad](./trulens_eval/core_concepts_rag_triad.md), and [Honest, Harmless and Helpful Evals](./trulens_eval/core_concepts_honest_harmless_helpful_evals.md). +Read more about the core concepts behind TruLens including [Feedback Functions](https://www.trulens.org/trulens_eval/core_concepts_feedback_functions/), [The RAG Triad](https://www.trulens.org/trulens_eval/core_concepts_rag_triad/), and [Honest, Harmless and Helpful Evals](https://www.trulens.org/trulens_eval/core_concepts_honest_harmless_helpful_evals/). ## TruLens in the development workflow @@ -30,4 +30,4 @@ Walk through how to instrument and evaluate a RAG built from scratch with TruLen ### 💡 Contributing -Interested in contributing? See our [contribution guide](https://github.com/truera/trulens/tree/main/trulens_eval/CONTRIBUTING.md) for more details. +Interested in contributing? See our [contribution guide](https://www.trulens.org/trulens_eval/CONTRIBUTING/) for more details. diff --git a/trulens_eval/CONTRIBUTING.md b/trulens_eval/CONTRIBUTING.md index e8642282c..836f6ea89 100644 --- a/trulens_eval/CONTRIBUTING.md +++ b/trulens_eval/CONTRIBUTING.md @@ -36,7 +36,7 @@ New contributors may want to start with issues tagged with good first issue. Please feel free to open an issue and/or assign an issue to yourself. ## 🎉 Add Usage Examples -If you have applied TruLens to track and evalaute a unique use-case, we would love your contribution in the form of an example notebook: e.g. [Evaluating Pinecone Configuration Choices on Downstream App Performance](https://github.com/truera/trulens/blob/main/trulens_eval/examples/vector-dbs/pinecone/constructing_optimal_pinecone.ipynb) +If you have applied TruLens to track and evalaute a unique use-case, we would love your contribution in the form of an example notebook: e.g. [Evaluating Pinecone Configuration Choices on Downstream App Performance](https://colab.research.google.com/github/truera/trulens/blob/main/trulens_eval/examples/expositional/vector-dbs/pinecone/pinecone_evals_build_better_rags.ipynb) All example notebooks are expected to: diff --git a/trulens_eval/README.md b/trulens_eval/README.md index 078f95ed8..fa4c99997 100644 --- a/trulens_eval/README.md +++ b/trulens_eval/README.md @@ -6,7 +6,7 @@ Fine-grained, stack-agnostic instrumentation and comprehensive evaluations help you to identify failure modes & systematically iterate to improve your application. -Read more about the core concepts behind TruLens including [Feedback Functions](./trulens_eval/core_concepts_feedback_functions.md), [The RAG Triad](./trulens_eval/core_concepts_rag_triad.md), and [Honest, Harmless and Helpful Evals](./trulens_eval/core_concepts_honest_harmless_helpful_evals.md). +Read more about the core concepts behind TruLens including [Feedback Functions](https://www.trulens.org/trulens_eval/core_concepts_feedback_functions/), [The RAG Triad](https://www.trulens.org/trulens_eval/core_concepts_rag_triad/), and [Honest, Harmless and Helpful Evals](https://www.trulens.org/trulens_eval/core_concepts_honest_harmless_helpful_evals/). ## TruLens in the development workflow @@ -30,4 +30,4 @@ Walk through how to instrument and evaluate a RAG built from scratch with TruLen ### 💡 Contributing -Interested in contributing? See our [contribution guide](https://github.com/truera/trulens/tree/main/trulens_eval/CONTRIBUTING.md) for more details. +Interested in contributing? See our [contribution guide](https://www.trulens.org/trulens_eval/CONTRIBUTING/) for more details.