Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[R-263] Roadmap - v0.2 #1009

Open
7 of 26 tasks
jjmachan opened this issue Jun 5, 2024 · 9 comments
Open
7 of 26 tasks

[R-263] Roadmap - v0.2 #1009

jjmachan opened this issue Jun 5, 2024 · 9 comments
Labels
enhancement New feature or request

Comments

@jjmachan
Copy link
Member

jjmachan commented Jun 5, 2024

image

From SyncLinear.com | R-263

@jjmachan jjmachan changed the title Roadmap - v0.2 [R-263] Roadmap - v0.2 Jun 5, 2024
@jjmachan jjmachan added the enhancement New feature or request label Jun 5, 2024
@jjmachan jjmachan modified the milestone: v0.1.9 Jun 5, 2024
@dosubot dosubot bot added the module-metrics this is part of metrics module label Jun 5, 2024
@jjmachan jjmachan pinned this issue Jun 5, 2024
@jjmachan jjmachan removed the module-metrics this is part of metrics module label Jun 6, 2024
@rajib76
Copy link

rajib76 commented Jun 6, 2024

We lack chunk quality metrics as of today. It will be good to see some chunk quality evaluation metrics.

@jjmachan
Copy link
Member Author

jjmachan commented Jun 8, 2024

hey @rajib76, thanks for chipping in 🙂

could you explain a bit more about how you're measuring quality here? maybe an example too if possible?

@rajib76
Copy link

rajib76 commented Jun 8, 2024

One of the hard problem today in RAG is to determine the right size of the chunk. If a chunk talks about multiple concept, it is very difficult to find the most relevant chunk for the question. I was looking for a metrics that will tell that a chunk is atomic and it talks about only one concept. The semantic chunking approach did not work as the embedding model itself has a semantic dissonance.

@sky-2002
Copy link
Contributor

@jjmachan @rajib76 metrics like chunk_attribution and chunk_utilization (as referenced here) could help to quantify chunk quality. We already have relevance scores(from vector DBs or keyword search engines) to measure chunk relevance with respect to the query. But metrics to quantify how much of the chunk was used can be helpful. I can can take this up you find them useful, I found it interesting, could help decide how many chunks to retrieve.

@rajib76
Copy link

rajib76 commented Aug 18, 2024 via email

@alexander-zuev
Copy link

@jjmachan hey, thanks for the superb library and effort!

I want to use RAGAS to evaluate my open-source RAG application that has it's own custom chunker and retriever. Do you consider feasible adding support for custom-chunks to the synthetic data generator?

Right now I can't really use ragas fully because I need to rely on the chunks generated by ragas, instead of my own chunker.

@jjmachan
Copy link
Member Author

jjmachan commented Oct 5, 2024

hey @Twist333d - thanks for the kind words ❤️
we just revamped the testset generation piece for v0.2 - we just released an beta version - but main version should be out next week

do you want to give that a go?

@alexander-zuev
Copy link

yep @jjmachan shoot it of course!

Btw, I've just setup RAGAS to be used with Weave, and another feature request came up - it would be great if you supported a much easier integration with tracing & eval suites such as Weave by W&B.

@alexander-zuev
Copy link

Several more feature requests:

  1. Allow random or 'smart' sampling of samples to be used to generate questions. For example, I want to generate a test dataset for a set of documents. Depending on the volume (1 page vs 10000 pages), I want to be able to control how / where do the questions come from
  2. Async / parallel generation of embeddings
  3. Control over how many embeddings are generated for the input file to the dataset generator - as I understand right now it converts all nodes to embeddings which might be too costly or not necessary. For example, if I set test_size==1, why does it convert all nodes to embeddings

@shahules786 shahules786 unpinned this issue Oct 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

4 participants