local_fire_departmentHoneystax
search⌘K
loginLog Inperson_addSign Up
layers
HONEYSTAX TERMINAL v1.0
HomeNewsSavedSubmit
Back to the live board
I

index

Agent

The SOTA Open-Source Browser Agent for autonomously performing complex tasks on the web

Copy the install, test the workflow, then decide if it earns a permanent slot.

2,341
Why nowArchived

Strong idea surface, weaker maintenance signal. Steal the pattern before you commit.

DecisionHigh-conviction move

Copy the install, test the workflow, then decide if it earns a permanent slot.

Trial costMedium lift

Reasonable to try, but it will take more than a quick skim to get real signal.

Risk100/100

GitHub health 50/100. no security policy. Archived repo plus maintenance drag means copy patterns, not the whole tool.

What You Are Adopting

AI Agent

Universal

Model

Multiple

Build Time

Minutes

Test This In Your Stack

One command inClean rollbackLow commitment
shieldSandboxedInstalls to ~/.claude — isolated from your projects. One command to remove.

Fastest way to find out if index belongs in your setup.

Copy the install command, run a real test, and back it out cleanly if it slows you down.

Try now
git clone https://github.com/lmnr-ai/index ~/.claude/agents/index

Run this first. You will know quickly if the workflow earns a permanent slot.

Back out
rm -rf ~/.claude/agents/index

No messy cleanup loop. If it misses, remove it and keep moving.

Install Location

~/  └─ .claude/      ├─ commands/      ├─ agents/      │   └─ index/ ← installs here      └─ settings.json

About

The SOTA Open-Source Browser Agent for autonomously performing complex tasks on the web. An open-source agent for the AI coding ecosystem.

README

GitHub stars Static Badge X (formerly Twitter) Follow Static Badge

Laminar logo

Index

Index is a state-of-the-art open-source browser agent that autonomously executes complex web tasks. It turns any website into an accessible API and can be seamlessly integrated with just a few lines of code.

  • Powered by reasoning LLMs with vision capabilities.
    • Gemini 2.5 Pro (really fast and accurate)
    • Claude 3.7 Sonnet with extended thinking (reliable and accurate)
    • OpenAI o4-mini (depending on the reasoning effort, provides good balance between speed, cost and accuracy)
    • Gemini 2.5 Flash (really fast, cheap, and good for less complex tasks)
  • pip install lmnr-index and use it in your project
  • index run to run the agent in the interactive CLI
  • Supports structured output via Pydantic schemas for reliable data extraction.
  • Index is also available as a serverless API.
  • You can also try out Index via Chat UI.
  • Supports advanced browser agent observability powered by open-source platform Laminar.

prompt: go to ycombinator.com. summarize first 3 companies in the W25 batch and make new spreadsheet in google sheets.

local_agent_spreadsheet_demo.mp4

Documentation

Check out full documentation here

Quickstart

Install dependencies

pip install lmnr-index 'lmnr[all]'

# Install playwright
playwright install chromium

Setup model API keys

Setup your model API keys in .env file in your project root:

GEMINI_API_KEY=
ANTHROPIC_API_KEY=
OPENAI_API_KEY=
# Optional, to trace the agent's actions and record browser session
LMNR_PROJECT_API_KEY=

Run Index with code

import asyncio
from index import Agent, GeminiProvider
from pydantic import BaseModel
from lmnr import Laminar
import os

# to trace the agent's actions and record browser session
Laminar.initialize()

# Define Pydantic schema for structured output
class NewsSummary(BaseModel):
    title: str
    summary: str

async def main():

    llm = GeminiProvider(model="gemini-2.5-pro-preview-05-06")
    agent = Agent(llm=llm)

    # Example of getting structured output
    output = await agent.run(
        prompt="Navigate to news.ycombinator.com, find a post about AI, extract its title and provide a concise summary.",
        output_model=NewsSummary
    )
    
    summary = NewsSummary.model_validate(output.result.content)
    print(f"Title: {summary.title}")
    print(f"Summary: {summary.summary}")
    
if __name__ == "__main__":
    asyncio.run(main())

Run Index with CLI

Index CLI features:

  • Browser state persistence between sessions
  • Follow-up messages with support for "give human control" action
  • Real-time streaming updates
  • Beautiful terminal UI using Textual

You can run Index CLI with the following command.

index run

Output will look like this:

Loaded existing browser state
╭───────────────────── Interactive Mode ─────────────────────╮
│ Index Browser Agent Interactive Mode                       │
│ Type your message and press Enter. The agent will respond. │
│ Press Ctrl+C to exit.                                      │
╰────────────────────────────────────────────────────────────╯

Choose an LLM model:
1. Gemini 2.5 Flash
2. Claude 3.7 Sonnet
3. OpenAI o4-mini
Select model [1/2] (1): 3
Using OpenAI model: o4-mini
Loaded existing browser state

Your message: go to lmnr.ai, summarize pricing page

Agent is working...
Step 1: Opening lmnr.ai
Step 2: Opening Pricing page
Step 3: Scrolling for more pricing details
Step 4: Scrolling back up to view pricing tiers
Step 5: Provided concise summary of the three pricing tiers

Running CLI with a personal Chrome instance

You can use Index with personal Chrome browser instance instead of launching a new browser. Main advantage is that all your existing logged-in sessions will be available.

# Basic usage with default Chrome path
index run --local-chrome

Use Index via API

The easiest way to use Index in production is with serverless API. Index API manages remote browser sessions, agent infrastructure and browser observability. To get started, create a project API key in Laminar.

Install Laminar

pip install lmnr

Use Index via API

from lmnr import Laminar, LaminarClient
# you can also set LMNR_PROJECT_API_KEY environment variable

# Initialize tracing
Laminar.initialize(project_api_key="your_api_key")

# Initialize the client
client = LaminarClient(project_api_key="your_api_key")

for chunk in client.agent.run(
    stream=True,
    model_provider="gemini",
    model="gemini-2.5-pro-preview-05-06",
    prompt="Navigate to news.ycombinator.com, find a post about AI, and summarize it"
):
    print(chunk)
    

Browser agent observability

Both code run and API run provide advanced browser observability. To trace Index agent's actions and record browser session you simply need to initialize Laminar tracing before running the agent.

from lmnr import Laminar

Laminar.initialize(project_api_key="your_api_key")

Then you will get full observability on the agent's actions synced with the browser session in the Laminar platform. Learn more about browser agent observability in the documentation.

Index observability

Made with ❤️ by the Laminar team

Tech Stack

GoOpenAIClaudeLLM

Installation

pip install lmnr

Open Live ProjectAudit Repo

Reviews0

Log in to write a review.

ArchivedThis repo is archived
bug_report6open issues
Submitted November 30, 2024

auto_awesomeYour strongest next moves after index