ZiminPark / coffee-augmented-rag

4 stars 0 forks source link

Hypothesis 부분 제대로 이해했는지 확인하기 #9

Closed ZiminPark closed 3 months ago

ZiminPark commented 3 months ago

We hypothesize that encoder-decoder models may make better use of their context windows because their bidirectional encoder allows processing each document in the context of future documents, potentially improving relative importance estimation between documents.

image
ZiminPark commented 3 months ago

We hypothesize that non-instruction fine-tuned language models learn to use these long contexts from similarlyformatted data that may occur in Internet text seen during pre-training, e.g., StackOverflow questions and answers.

image