I Forked Karpathy's LLM Council project... The Result is INSANE! πŸš€

I Forked Karpathy's LLM Council project... The Result is INSANE! πŸš€

Introduction to User Experience Project

Overview of the Project

  • Jacob introduces a project focused on user experience, inspired by Andre Kapathy's concept of "vibe coding."
  • The project utilizes a tool called LLM Council, which allows users to query multiple language models (LLMs) and receive peer-reviewed answers.

Functionality of LLM Council

  • The original tool requires installation and configuration, making it less user-friendly for casual experimentation.
  • Jacob demonstrates the user interface where multiple models provide their answers to a question, followed by peer ranking.

Enhancements Made: Search Capability and More

Initial Limitations Observed

  • Jacob notes that responses from the models were limited due to their training data cutoffs, particularly regarding current events like the AI bubble.

Upgrades Implemented

  • He aimed to enhance the tool with search capabilities but ended up adding more features through what he calls "vibe product management" or VIPMing.

Introducing LLM Council Plus

New Features Added

  • Jacob presents his fork of the original tool named "LLM Council Plus," which includes additional functionalities beyond the original version.

Visual Enhancements

  • A visual representation shows participants in the council, including a designated chairman for decision-making.

User-Friendly Configuration Options

API Key Management

  • A settings page has been added for easier management of LLM API keys without needing complex configurations.

Local Model Support

  • Users can now utilize local models alongside paid APIs, allowing for side-by-side comparisons during testing.

Advanced Customization Features

System Prompt Control

  • The application allows customization of system prompts sent to models, enabling varied response styles based on user-defined parameters.

Creativity Controls

  • Users can adjust creativity levels (temperature settings), influencing how conservative or creative model responses are.

User Control and Search Features in AI Models

Overview of User Control

  • The interface allows users to have full control over various stages, including a reset option to revert settings to default.
  • Multiple search engines can be integrated, such as DuckDuckGo (limited free access), Tavi (requires API keys), Brave Search, and Gina AI for article retrieval.

Challenges with Information Retrieval

  • Some websites block certain services, leading to truncated responses; however, many sites still provide well-formatted AI responses.
  • Users can create unique councils with pre-selected models that can be exported or imported for future use.

Configuration and Model Selection

  • Users can select specific providers like OpenRouter and filter for free models while being warned about potential rate limits when using multiple free models from one provider.
  • There is a batching feature for adding multiple council members but caution is advised due to possible errors.

Interactive Features in the New Chat Interface

Enhanced Interaction Capabilities

  • A new feature allows users to enable or disable search functionality during chats, providing flexibility in interaction.
  • Users can ask questions across multiple models without ratings initially, allowing for unbiased comparisons of responses.

Peer Ranking System

  • After receiving answers, there is a peer ranking stage where each model's response is evaluated against others.
  • A visual leaderboard displays the rankings of different models based on their performance in answering queries.

Future Enhancements and Ideas

  • The speaker expresses interest in developing a local arena for ongoing queries but acknowledges time constraints limiting further development.
Video description

Did Andrej Karpathy's excellent "LLM Council" feel too raw and limited? I fixed it. In this video, I reveal "LLM Council Plus" - a fully upgraded fork that adds a professional UI, web search capabilities, and full support for ANY LLMs via APIs as local models through Ollama. ABOUT THIS VIDEO: The original LLM Council was a brilliant concept, but it lacked usability and choice. I spent a couple of weekends improving it and adding tons of options. I cover how to connect OpenRouter, run local Ollama models like Phi-4, and use a real-time web search tool. KEY TAKEAWAYS: βœ… Run Local LLM Councils using Ollama (Privacy-focused) and remote AI providers βœ… Compare 8+ Models simultaneously with a Visual Leaderboard βœ… Enable Web Search for real-time fact-checking CHAPTERS: 00:00 - The Problem with Raw Python Scripts 02:05 - Introducing LLM Council Plus (The Upgrade) 03:20 - Adding Local Models (Ollama Support) 05:33 - Customizing System Prompts & Temperature 06:50 - Enabling Web Search (DuckDuckGo & Jina) 08:08 - The "I'm Feeling Lucky" Model Selector 10:35 - Live Demo: Full Council Deliberation 11:45 - The Final Verdict & Rankings 12:02 - How to Get the Code RESOURCES & LINKS: πŸ”— Get the Code (GitHub): https://github.com/jacob-bd/llm-council-plus πŸ”— Original Repo (Karpathy): https://github.com/karpathy/llm-council πŸ”” Subscribe for more cool AI projects and Vibe Coding tips. #LLMCouncil #AndrejKarpathy #Ollama #LocalLLM #VibeCoding #VibePMing