Closed writinwaters closed 1 month ago
INFO:werkzeug:WARNING: This is a development server. Do not use it in a production deployment. Use a production WSGI server instead.
v0.1.0
- [x] URL support: Capable of web crawling and the corresponding content extration. @KevinHuSh
@KevinHuSh Hi, I wonder what's the current status? Maybe i can collaborate on it. My tech Stack is python backend and i think web crawling is essential for workflow of software development.
v0.1.0
- [x] URL support: Capable of web crawling and the corresponding content extration. @KevinHuSh
@KevinHuSh Hi, I wonder what's the current status? Maybe i can collaborate on it. My tech Stack is python backend and i think web crawling is essential for workflow of software development.
Crawling web page is big thing in my understanding. We do not have a clear picture for this. Here are two important points to note: Crawling Task Dispaching The way to execute JS on the page Page classification which is related the structure of the data we store The extraction of the main parts of the page
If you have any good solution to these points, please let me know....
Crawling web page is big thing in my understanding. We do not have a clear picture for this. Here are two important points to note: Crawling Task Dispaching The way to execute JS on the page Page classification which is related the structure of the data we store The extraction of the main parts of the page
If you have any good solution to these points, please let me know....
@KevinHuSh Please refer to #315
Maybe i can start with AWS Bedrock models
to contribute to the project, then Support x-inference as model provider
. Feel free to contact me via here or wx.
@writinwaters @KevinHuSh Hi. Requesting to look at the issue I created: https://github.com/infiniflow/ragflow/issues/345
Maybe fixing these issues would help us adopt ragflow better.
Not supporting streaming really affects the user experience. I hope it can be supported soon, as the implementation is not complicated.
v0.6.0
- [ ] Print version or commit-id when RAGFlow is started. Or showing these information on UI. [Feature Request]: Print version or commit-id when RAGFlow is started. #643
- [ ] APIs for knowledge base and file manager. [Feature Request]: File manager & API #345
- [ ] Files in knowledge base should also be found in file manager. let file in knowledgebases visible in file manager #714
- [ ] System components monitoring.
- [ ] Supports simple document layout to speed up file parsing.
- [ ] Supports user configured reranker.
- [ ] Streaming conversation output. [Feature Request]: Support for conversational streaming #709
Whether it can provide users with accurate answers and quick answers, one is subject to accuracy and the other is subject to quick response
v0.6.0
- [ ] Print version or commit-id when RAGFlow is started. Or showing these information on UI. [Feature Request]: Print version or commit-id when RAGFlow is started. #643
- [ ] APIs for knowledge base and file manager. [Feature Request]: File manager & API #345
- [ ] Files in knowledge base should also be found in file manager. let file in knowledgebases visible in file manager #714
- [ ] System components monitoring.
- [ ] Supports simple document layout to speed up file parsing.
- [ ] Supports user configured reranker.
- [ ] Streaming conversation output. [Feature Request]: Support for conversational streaming #709
Whether it can provide users with accurate answers and quick answers, one is subject to accuracy and the other is subject to quick response
You can file a new issue, so we can discuss in that issue.
Are there any plans to use big language models for knowledge graphs?
Are there any plans to use big language models for knowledge graphs?
We've been thinking about this for a while, but haven't figured out how to implement it in RAGFlow. If you have any good issues, feel free to create a new issue and we'll discuss it!
Can it automatically continue when it says 'Due to length...' ? The current handling of length issues feels very rudimentary.
Postpone feature request of reranker configuration to 0.7.0.
v0.1.0
- [x] URL support: Capable of web crawling and the corresponding content extration. @KevinHuSh
@KevinHuSh Hi, I wonder what's the current status? Maybe i can collaborate on it. My tech Stack is python backend and i think web crawling is essential for workflow of software development.
Crawling web page is big thing in my understanding. We do not have a clear picture for this. Here are two important points to note: Crawling Task Dispaching The way to execute JS on the page Page classification which is related the structure of the data we store The extraction of the main parts of the page
If you have any good solution to these points, please let me know....
There is already an existing service that does this job pretty well, could be interesting to look into this for crawling website content: https://www.firecrawl.dev/
v0.1.0
- [x] URL support: Capable of web crawling and the corresponding content extration. @KevinHuSh
@KevinHuSh Hi, I wonder what's the current status? Maybe i can collaborate on it. My tech Stack is python backend and i think web crawling is essential for workflow of software development.
Crawling web page is big thing in my understanding. We do not have a clear picture for this. Here are two important points to note: Crawling Task Dispaching The way to execute JS on the page Page classification which is related the structure of the data we store The extraction of the main parts of the page If you have any good solution to these points, please let me know....
There is already an existing service that does this job pretty well, could be interesting to look into this for crawling website content: https://www.firecrawl.dev/
We will look into this website.
v0.10.0
v0.9.0
v0.8.0
v0.7.0
v0.6.0
Long-term plan