How does an AI application builder choose a language model?
To choose a model off the market today, companies perform evaluations on their private data sets. The greatest difference in evaluation is made by the quality of the evaluation dataset which is very closely tied to the task the models are used for.
1. Of the 270 odd languages in India- where do we begin?
2. What are the important domains to be looking at? And what use cases are most relevant to us?
1. What are the use cases and language model tasks that we should evaluate for?
2. What factors are most important to you when choosing a model - for example, open-source availability, multilingual capabilities, or specific performance metrics? How do you prioritise these factors?
3. How do you define and measure "success" for the tasks where you're using language models? What metrics or outcomes are most critical ?
4. How do you balance trade-offs between different metrics for various tasks?
5. How can data sets be maintained and enriched by private evaluations?
How you can contribute
1. Join our group that believes in the goal of generating Ten Trillion Indic tokens, and approach partners.
2. Pick a project that resonates with you
3. Fund the project and/or suggest incentive schemes