diff --git a/docs/getting_started/customization.rst b/docs/getting_started/customization.rst index f0a152d2b1f2e500486560135c58e10cb9514995..3ca772a365a3336a9a53e31dfe2cdd45f980032e 100644 --- a/docs/getting_started/customization.rst +++ b/docs/getting_started/customization.rst @@ -25,7 +25,7 @@ In this tutorial, we start with the code you wrote for the `starter example <sta service_context = ServiceContext.from_defaults(chunk_size=1000) -The `ServiceContext </module_guides/supporting_modules/service_context.html>`_ is a bundle of services and configurations used across a LlamaIndex pipeline. +The `ServiceContext <../module_guides/supporting_modules/service_context.html>`_ is a bundle of services and configurations used across a LlamaIndex pipeline. .. code-block:: python :emphasize-lines: 4 @@ -53,7 +53,7 @@ The `ServiceContext </module_guides/supporting_modules/service_context.html>`_ i vector_store = ChromaVectorStore(chroma_collection=chroma_collection) storage_context = StorageContext.from_defaults(vector_store=vector_store) -`StorageContext` defines the storage backend for where the documents, embeddings, and indexes are stored. You can learn more about `storage </module_guides/storing/storing.html>`_ and `how to customize it </module_guides/storing/customization.html>`_. +`StorageContext` defines the storage backend for where the documents, embeddings, and indexes are stored. You can learn more about `storage <../module_guides/storing/storing.html>`_ and `how to customize it <../module_guides/storing/customization.html>`_. .. code-block:: python :emphasize-lines: 4 @@ -81,7 +81,7 @@ The `ServiceContext </module_guides/supporting_modules/service_context.html>`_ i response = query_engine.query("What did the author do growing up?") print(response) -`as_query_engine` builds a default `retriever` and `query engine` on top of the index. You can configure the retriever and query engine by passing in keyword arguments. Here, we configure the retriever to return the top 5 most similar documents (instead of the default of 2). You can learn more about `retrievers </module_guides/querying/retrievers.html>_` and `query engines </module_guides/putting_it_all_together/query_engine/root.html>_` +`as_query_engine` builds a default `retriever` and `query engine` on top of the index. You can configure the retriever and query engine by passing in keyword arguments. Here, we configure the retriever to return the top 5 most similar documents (instead of the default of 2). You can learn more about `retrievers <../module_guides/querying/retrievers.html>_` and `query engines <../module_guides/putting_it_all_together/query_engine/root.html>_` ----------------- @@ -94,7 +94,7 @@ The `ServiceContext </module_guides/supporting_modules/service_context.html>`_ i service_context = ServiceContext.from_defaults(llm=PaLM()) -You can learn more about `customizing LLMs </module_guides/models/llms.html>`_. +You can learn more about `customizing LLMs <../module_guides/models/llms.html>`_. .. code-block:: python :emphasize-lines: 5 @@ -123,7 +123,7 @@ You can learn more about `customizing LLMs </module_guides/models/llms.html>`_. response = query_engine.query("What did the author do growing up?") print(response) -You can learn more about `query engines <../core_modules/query_modules/query_engine/usage_pattern.html>`_ and `response modes </module_guides/putting_it_all_together/query_engine/response_modes.html>`_. +You can learn more about `query engines <../module_guides/querying/querying.html>`_ and `response modes <../module_guides/putting_it_all_together/query_engine/response_modes.html>`_. ----------------- @@ -141,7 +141,7 @@ You can learn more about `query engines <../core_modules/query_modules/query_eng response = query_engine.query("What did the author do growing up?") response.print_response_stream() -You can learn more about `streaming responses </module_guides/putting_it_all_together/query_engine/streaming.html>`_. +You can learn more about `streaming responses <../module_guides/putting_it_all_together/query_engine/streaming.html>`_. ----------------- @@ -161,11 +161,11 @@ You can learn more about `streaming responses </module_guides/putting_it_all_tog response = query_engine.chat("Oh interesting, tell me more.") print(response) -Learn more about the `chat engine </module_guides/putting_it_all_together/chat_engines/usage_pattern.html>`_. +Learn more about the `chat engine <../module_guides/putting_it_all_together/chat_engines/usage_pattern.html>`_. ----------------- .. admonition:: Next Steps - * want a thorough walkthrough of (almost) everything you can configure? Get started with `Understanding LlamaIndex </understanding/understanding.html>`_. + * want a thorough walkthrough of (almost) everything you can configure? Get started with `Understanding LlamaIndex <../understanding/understanding.html>`_. * want more in-depth understanding of specific modules? Check out the module guides in the left nav π diff --git a/docs/index.rst b/docs/index.rst index 204b27fcda6d84341e32938222f0a710c5fe6925..0daa81e995bdb9f317c0bb381e56538e649607d0 100644 --- a/docs/index.rst +++ b/docs/index.rst @@ -41,7 +41,7 @@ To install the library: ``pip install llama-index`` -We recommend starting at `how to read these docs <getting_started/reading.md>`_, which will point you to the right place based on your experience level. +We recommend starting at `how to read these docs <./getting_started/reading.md>`_, which will point you to the right place based on your experience level. πΊοΈ Ecosystem ************