Talk to Your Documents, Powered by Llama-Index

Talk to Your Documents, Powered by Llama-Index

Prompt Engineering

1 год назад

86,671 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

@trobinsun9851
@trobinsun9851 - 26.09.2023 21:27

what about security of our data ? what if it's confidential document ? thanks for your excellent videos

Ответить
@y2knoproblem
@y2knoproblem - 27.09.2023 03:33

This is very helpful. Where can we find the system architecture diagram?

Ответить
@gitinit3416
@gitinit3416 - 27.09.2023 07:19

Awesome ❤ a little off topic question......would be so kind as to share the app you are using for making diagrams. it's sick....I've been looking for something that that since quite a while now, but with no luck.... 🥺

Ответить
@sachinkalsi9516
@sachinkalsi9516 - 27.09.2023 18:36

Thanks for the video. I’m facing lots of latency issues (15+ min) while reading the stores the index. How I can improve ? There are 100k+ vectors . Going ahead with numpy array takes few minutes only !

Ответить
@ChronozOdP
@ChronozOdP - 28.09.2023 13:13

Another excellent video. Easy to follow and up to date. Thank you and keep it up!

Ответить
@nickwoolley733
@nickwoolley733 - 29.09.2023 00:47

Will you be testing the new Mistral-7B-v0.1 and Mistral-7B-Instruct-v0.1 LLMs? They claim to outperform Llama 2.😊

Ответить
@hernandocastroarana6206
@hernandocastroarana6206 - 29.09.2023 06:05

Excellent video. Do you know what is the best option to start the code in an interface? I passed the code to Vs Code and then started it in Streamlit but it gives me some problems. I appreciate your help

Ответить
@DixitNitish
@DixitNitish - 29.09.2023 17:38

Great video, how can I have the ability to compare 100s of document using llamaindex and will it know which chunk belongs to which document when answering the question? Also, how do you make sure all the pieces that should be in 1 chunk stays together, for instance if there is a table that goes across 2 pages then that should still be in 1 chunk?

Ответить
@adamduncan6579
@adamduncan6579 - 29.09.2023 17:53

Excellent videos! Really helping out with my work. Curious what tool you are using to draw the system architecture? I really like the way it renders the architectures.

Ответить
@matten_zero
@matten_zero - 30.09.2023 13:04

Much more intuitive than LangChain

Ответить
@henkhbit5748
@henkhbit5748 - 30.09.2023 21:45

Nice intro about llma-index👍. I think for small amount of documents the default llma-index embedding in json is sufficient. I suppose u can also use chromadb or weaviate or other vectorstores. Would be nice to see with the non default vector store...

Ответить
@KinoInsight
@KinoInsight - 02.10.2023 07:36

I liked your explanation. You are a good story teller. You explained the details in a simple way and yet easy to implement manner. Thank you. I look forward to your next video.

But how do we ensure the latest data is fed to the LLM in real time? In this case, we need to provide the data to the Llama. And the response is limited to the data provided.

Ответить
@anuvratshukla7061
@anuvratshukla7061 - 07.10.2023 10:39

If I'm uusing weaviate, how to load then?

Ответить
@vladimirgorea8714
@vladimirgorea8714 - 11.10.2023 04:09

what's the difference in using Llamaindex and just using openai embeddings?

Ответить
@nazihfattal974
@nazihfattal974 - 13.10.2023 08:55

I tried this in my colab pro account and the session crashed when I ran the vectorstore.
Out of GPU memory. colab allocated 16GB of VRAM.

Would you please add option for using huggingface hosted LLMs through their free inference API (applies to select models)?
Thanks for a great video.

Ответить
@hamtsammich
@hamtsammich - 20.10.2023 01:21

could you do a tutorial about how to do this locally?
I'm very interested in llama index, but I'm wary of using things that aren't on my local hardware

Ответить
@xt3708
@xt3708 - 20.10.2023 15:06

learn so much from you!

Ответить
@CaesarEduBiz-lz2cg
@CaesarEduBiz-lz2cg - 22.10.2023 18:52

Is it better to use Llama Index or RAG (Retrieval Augmented Generation) ?

Ответить
@am1rsafavi-naini356
@am1rsafavi-naini356 - 23.10.2023 13:43

I don't have any credit card, but I will buy an coffee for you some day (maybe in person, who knows :)

Ответить
@Rahul-zq8ep
@Rahul-zq8ep - 07.11.2023 13:43

is it a Production ready code ? What important points we should keep in mind to make a similar app for Production environment ?

Ответить
@ismaelnoble
@ismaelnoble - 08.11.2023 23:54

how does llama index compare with the localgpt method?

Ответить
@aseemasthana4121
@aseemasthana4121 - 14.11.2023 18:45

Perfect pace and level of knowledge. Loved the video.

Ответить
@huyvo9105
@huyvo9105 - 03.12.2023 05:20

Hi, can i get your pipeline draw link please

Ответить
@abdullahiahmad3244
@abdullahiahmad3244 - 06.12.2023 01:27

to be honest this is the best tutorial i see in 2023

Ответить
@s.moneebahnoman
@s.moneebahnoman - 09.12.2023 20:32

Amazing! I haven't seen enough videos talking about persisting the index especially in beginner level tutorials. I think its such a crucial concept that I found out much later. Love the flow for this and its perfectly explained! Liked and subbed!

Ответить
@zearcher4633
@zearcher4633 - 10.12.2023 21:50

can you make a video on how to create a website chatbot out of all of this? say, we used this video and made a chatbot to talk with our data, how do we use it in our website?

Ответить
@Udayanverma
@Udayanverma - 12.12.2023 06:07

Do u have any example of a model on personal desktop/server. I dont wish to publish my content to chatgpt or any internet service.

Ответить
@nishkarve
@nishkarve - 26.12.2023 01:46

Excellent. Is there a video you are planning to make on a multi modal RAG? I have a PDF which is an instruction manual. It has text and images. When a user asks a question, for example, "How to connect the TV to external speakers?", it should show the steps and the images associated with those steps. Everywhere I see are examples of image "generation". I don't want to generate images. I just want to show what's in the PDF based on the question.

Ответить
@dimitripetrenko438
@dimitripetrenko438 - 29.12.2023 20:13

Hi bro cool video! May I ask if there is a way to store quantized model with LlamaIndex? It's very painful to quantize it every single time I try to run it

Ответить
@mohsenghafari7652
@mohsenghafari7652 - 04.01.2024 10:23

hi. googlecolab link give error

Ответить
@jamesvictor2182
@jamesvictor2182 - 19.01.2024 16:54

Great video. The notebook fails at the first hurdle for me: ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
llmx 0.0.15a0 requires cohere, which is not installed.
tensorflow-probability 0.22.0 requires typing-extensions<4.6.0, but you have typing-extensions 4.9.0 which is incompatible.

Ответить
@dineshbhatotia8783
@dineshbhatotia8783 - 27.01.2024 23:17

ValueError: The current `device_map` had weights offloaded to the disk. Please provide an `offload_folder` for them. Alternatively, make sure you have `safetensors` installed if the model you are using offers the weights in this format.

Getting this error

Ответить
@udithweerasinghe6402
@udithweerasinghe6402 - 30.01.2024 04:57

Do you have codes to do the same without openAI. Using some model in huggingface?

Ответить
@rizwanat7496
@rizwanat7496 - 03.02.2024 20:15

I am in the vectorstoreindex.from documents cell. it been running for like 24 hrs now. How do I know when it will end. I am running it locally in my laptop. output shows completion of batches. almost 2400+ batches. but it doesn;t showing how many are left. can somebody help?. my data consist of 850+ json. over all 70MB data.

Ответить
@sidhaarthsredharan3318
@sidhaarthsredharan3318 - 08.03.2024 17:06

im doing the same, but indexing every node created, there are around 5000 nodes, and its taking a long time. is there some progress bar (like tqdm) code i can add to see how long the indexing process would take?

Ответить
@elizonfrankcarcaustomamani4999
@elizonfrankcarcaustomamani4999 - 11.03.2024 03:30

Hello, help with that please. When I execute the line 'index = VectorStoreIndex.from_documents(documents)' after 1 min I get an error 429 (insufficient_quota). Check if the OPENAI_API_KEY variable was registered with '!export -p', and if it is. Thanks

Ответить
@MercyBai-i2t
@MercyBai-i2t - 14.03.2024 17:26

Can I build a multilingual chatbot using llama index?

Ответить
@AI_Expert_007
@AI_Expert_007 - 25.04.2024 08:41

Thanks for the clear explanation. Could you please share the name of the tool you used to create the workflow diagram?

Ответить
@kayasaz
@kayasaz - 31.05.2024 10:16

Great explanation

Ответить
@dario27
@dario27 - 27.06.2024 16:55

Finally a good tutorial on the subject! Thanks so much!

Ответить
@HarmeetSingh-ry6fm
@HarmeetSingh-ry6fm - 28.06.2024 15:41

Hi prompt as you mentioned in this video that this is a system prompt for StableLM, I want to know is there a way I can find prompt format for different LLM for example mixtral 8x7b/22b or llama 3

Ответить
@sanjaybhatikar
@sanjaybhatikar - 27.08.2024 11:37

Clear and concise explanations

Ответить
@sanjaybhatikar
@sanjaybhatikar - 30.08.2024 12:55

Nice video! Dependency conflicts are legion with both LangChain and LlamaIndex. They have rushed tools to production but sotware quality has been sacrificed for speed. That trend is here to stay - we spend more time resolving dependency conflicts than coding our apps. Your videos are very good.

Ответить
@superfreiheit1
@superfreiheit1 - 01.11.2024 17:45

Can you show how to do it with free and local llms

Ответить
@PickMyNick
@PickMyNick - 13.11.2024 00:40

Hello! Nice intro and diagram, quite ok intro do concepts. Unfortunately number of stuff is already deprecated, obsolete, changed in API with breaking way, so no chance to use the colab without digging or looking at current api contexts :(

Ответить