Azure-Samples / azure-search-openai-demo

A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure AI Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.
https://azure.microsoft.com/products/search
MIT License
5.57k stars 3.74k forks source link

How i automatize replacing of a file/document with a new version ? #1726

Open nitedan opened 1 week ago

nitedan commented 1 week ago

Please provide us with the following information:

This issue is for a: (mark with an x)

- [ ] bug report -> please search issues before submitting
- [ ] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)

Minimal steps to reproduce

Any log messages given by the failure

Expected/desired behavior

OS and Version?

Windows 7, 8 or 10. Linux (which distribution). macOS (Yosemite? El Capitan? Sierra?)

azd version?

run azd version and copy paste here.

Versions

Mention any other details that might be useful


Thanks! We'll be in touch soon.

pamelafox commented 1 week ago

The prepdocs script can be used with modifiers to remove a file and add a new file, per the documentation here: https://github.com/Azure-Samples/azure-search-openai-demo/blob/main/docs/data_ingestion.md#indexing-additional-documents

Can you specify what you mean by automate?

pamelafox commented 1 week ago

You might also look into integrated vectorization if you're looking for scheduled ingestion. That's also detailed in the same guide.

bp3000bp commented 1 day ago

may be a dumb or obvious question are there any additional costs that come along with scheduling ingestion for say every 2 hours?

berkyilmazer95 commented 21 hours ago

Hi, I read the following part from the guide. To add additional documents to the index, first upload them to your data source (Blob storage, by default). Then navigate to the Azure portal, find the index, and run it. The Azure AI Search indexer will identify the new documents and ingest them into the index.

But I still couldn't do it. :( I uploaded my own data into the blob. Then I opened it to the Search service via the portal. How can I run the current index from the index page?

Isn't an indexer needed to update the existing index? When I want to create an indexer for the current index, what skillset should I create before this? (I added the relevant blob in the Data sources tab in the Search service.)