AI Solutions
Discover and compare the best AI tools, rated by the community
Discover and compare the best AI tools, rated by the community
Roadmaps featuring essential concepts, learning methods, and the tools to put them into practice.
Curated List of AI Apps for productivity
AI Animal Explorer is an Omniverse extension that enables creators to quickly prototype unique 3D animal meshes.
AI tool from awesome-llm
The in-person certificate courses are not free, but all of the content is available on Fast.ai as MOOCs.
robust introduction to the subject and also the foundation for a Data Analyst “nanodegree” certification sponsored by Facebook and MongoDB.
AI tool from awesome-llm
A pioneering benchmark specifically designed to assess honesty in LLMs comprehensively.
evaluates LLM's ability to call external functions/tools.
AI tool from awesome-llm
AI tool from awesome-llm
AI tool from awesome-llm
AI tool from awesome-llm
A Challenging, Contamination-Free LLM Benchmark.
AI tool from awesome-llm
An Automatic Evaluator for Instruction-following Language Models using Nous benchmark suite.
AI tool from awesome-llm
a benchmark designed to evaluate large language models in the legal domain.
a benchmark designed to evaluate large language models (LLMs) specifically in their ability to answer real-world coding-related questions.
a benchmark evaluating QA methods that operate over a mixture of heterogeneous input sources (KB, text, tables, infoboxes).
a comprehensive benchmarking platform designed to evaluate large models' mathematical abilities across 20 fields and nearly 30,000 math problems.
CompassRank is dedicated to exploring the most advanced language and visual models, offering a comprehensive, objective, and neutral evaluation reference for the industry and research.
a ground-truth-based dynamic benchmark derived from off-the-shelf benchmark mixtures, which evaluates LLMs with a highly capable model ranking (i.e., 0.96 correlation with Chatbot Arena) while running locally and quickly (6% the time and cost of running MMLU).
a benchmark that evaluates large language models on a variety of multimodal reasoning tasks, including language, natural and social sciences, physical and social commonsense, temporal reasoning, algebra, and geometry.