With the introduction of LlamaIndex v0.10.0, there were several changes
With the introduction of LlamaIndex v0.10.0, there were several changes
- integrations have separate `pip installs (See the [full registry](https://pretty-sodium-5e0.notion.site/ce81b247649a44e4b6b35dfb24af28a6?v=53b3c2ced7bb4c9996b81b83c9f01139))
- integrations have separate `pip install`s (See the [full registry](https://pretty-sodium-5e0.notion.site/ce81b247649a44e4b6b35dfb24af28a6?v=53b3c2ced7bb4c9996b81b83c9f01139))
- many imports changed
- many imports changed
- the service context was deprecated
- the `ServiceContext` was deprecated
Thankfully, we've tried to make these changes as easy as possible!
Thankfully, we've tried to make these changes as easy as possible!
...
@@ -72,7 +72,7 @@ from llama_index.core import Settings
...
@@ -72,7 +72,7 @@ from llama_index.core import Settings
Settings.llm = llm
Settings.llm = llm
Settings.embed_model = embed_model
Settings.embed_model = embed_model
Setting.chunk_size = 512
Settings.chunk_size = 512
```
```
You can see the `ServiceContext` -> `Settings` migration guide for [more details](/module_guides/supporting_modules/service_context_migration.md).
You can see the `ServiceContext` -> `Settings` migration guide for [more details](/module_guides/supporting_modules/service_context_migration.md).
@@ -19,7 +19,7 @@ You may choose to **fine-tune** a LLM with your data, but:
...
@@ -19,7 +19,7 @@ You may choose to **fine-tune** a LLM with your data, but:
- Due to the cost to train, it's **hard to update** a LLM with latest information.
- Due to the cost to train, it's **hard to update** a LLM with latest information.
- **Observability** is lacking. When you ask a LLM a question, it's not obvious how the LLM arrived at its answer.
- **Observability** is lacking. When you ask a LLM a question, it's not obvious how the LLM arrived at its answer.
Instead of fine-tuning, one can a context augmentation pattern called `Retrieval-Augmented Generation (RAG) <./getting_started/concepts.html>`_ to obtain more accurate text generation relevant to your specific data. RAG involves the following high level steps:
Instead of fine-tuning, one can use a context augmentation pattern called `Retrieval-Augmented Generation (RAG) <./getting_started/concepts.html>`_ to obtain more accurate text generation relevant to your specific data. RAG involves the following high level steps:
1. Retrieve information from your data sources first,
1. Retrieve information from your data sources first,
2. Add it to your question as context, and
2. Add it to your question as context, and
...
@@ -36,7 +36,7 @@ In doing so, RAG overcomes all three weaknesses of the fine-tuning approach:
...
@@ -36,7 +36,7 @@ In doing so, RAG overcomes all three weaknesses of the fine-tuning approach:
Firstly, LlamaIndex imposes no restriction on how you use LLMs. You can still use LLMs as auto-complete, chatbots, semi-autonomous agents, and more (see Use Cases on the left). It only makes LLMs more relevant to you.
Firstly, LlamaIndex imposes no restriction on how you use LLMs. You can still use LLMs as auto-complete, chatbots, semi-autonomous agents, and more (see Use Cases on the left). It only makes LLMs more relevant to you.
LlamaIndex provides the following tools to help you quickly standup production-ready RAG systems:
LlamaIndex provides the following tools to help you quickly standup production-ready RAG systems:
- **Data connectors** ingest your existing data from their native source and format. These could be APIs, PDFs, SQL, and (much) more.
- **Data connectors** ingest your existing data from their native source and format. These could be APIs, PDFs, SQL, and (much) more.
- **Data indexes** structure your data in intermediate representations that are easy and performant for LLMs to consume.
- **Data indexes** structure your data in intermediate representations that are easy and performant for LLMs to consume.
...
@@ -70,7 +70,7 @@ We recommend starting at `how to read these docs <./getting_started/reading.html
...
@@ -70,7 +70,7 @@ We recommend starting at `how to read these docs <./getting_started/reading.html