Lionscraper MCP + CLI + HTTP API Bridge
L

Lionscraper MCP + CLI + HTTP API Bridge

LionScraper is a browser extension that enables web data collection through MCP, CLI, and HTTP API bridging tools.
2 points
7.3K

What is LionScraper?

LionScraper is an MCP (Model Context Protocol) server配套 with a browser extension. Its core function is to act as a 'bridge' to connect your AI application (such as Cursor) with the browser extension. You only need to give instructions in natural language in the AI application, and the MCP server will transmit these instructions to the browser extension, thereby achieving automated collection of web data, such as collecting lists, articles, links, images, email addresses, and phone numbers. It is like the 'eyes and hands' of an AI assistant, helping the AI obtain specific information you need from the web.

How to use LionScraper?

Using LionScraper mainly consists of three steps: 1. Basic preparation: Install the Chrome or Edge browser on your computer and install the LionScraper extension from the official store. 2. Deploy the server: Install the Node.js or Python runtime environment on your computer, then install the LionScraper package via npm or pip, and start its 'daemon'. 3. Configure the connection: Point the MCP configuration of your AI application (such as Cursor) to the installed LionScraper, and ensure that the 'Bridge Port' in the browser extension settings is consistent with the server configuration. After completing these steps, you can use natural conversations in the AI application to scrape web content.

Applicable scenarios

LionScraper is very suitable for scenarios that require batch extraction of structured data from web pages. For example: - During market research, collect product lists and prices from competitors' websites. - During content creation, screen and organize article titles and links from multiple news websites. - During data research, export table information from web pages with public data. - During daily office work, quickly scrape contact information (such as email addresses and phone numbers) from web pages. It greatly lowers the threshold of traditional web crawlers through natural language interaction, allowing users who are not familiar with programming to easily complete web data collection tasks.

Main Features

Triple Interface Support
It provides three access methods: MCP (for AI applications), CLI (command-line tool), and HTTP API, meeting the needs of different users and technology stacks. Whether you are using an AI programming assistant, are used to operating in the terminal, or want to integrate it into your own script, you can find a suitable way.
Natural Language-Driven Collection
Cross-Runtime Environment Support
It supports both Node.js and Python runtime environments. Users can choose to install the npm package or PyPI package according to their preferences or project requirements. Both provide the same functions and command-line interfaces.
Multiple Content Collection Types
It can intelligently collect various types of data from web pages, including lists, article bodies, hyperlinks, image URLs, email addresses, and phone numbers, basically covering most of the daily web data extraction needs.
HTTP Fallback Mode
When unable to connect to the Chrome or Edge browser and its extension, the MCP server can still start and use the 'http_fetch' mode to perform basic server-side HTTP GET requests, ensuring that basic functions are available while prompting users about the more powerful browser extension mode.
Background Daemon Mode
The CLI tool provides the `daemon` command, which can run as a background service (daemon) and continuously listen on a port, facilitating long-term collaborative work with the browser extension or other HTTP clients.
Advantages
Using natural language interaction, the operation threshold is low. Even users who don't understand complex technical commands can easily perform web data collection.
It has a rich variety of interface types, including MCP, CLI, and HTTP API, which can adapt to various usage scenarios such as AI applications, terminals, and programming scripts.
It has good cross-platform compatibility, providing both Node.js and Python installation packages, which can seamlessly integrate into the user's existing technical environment.
It has an HTTP fallback mode. Even when the browser is unavailable or the extension is not connected, the core functions are still available, ensuring the robustness of the system.
The scraping function is implemented by the browser extension, which can execute JavaScript and handle modern dynamic web pages, with stronger capabilities than pure server-side scraping.
Limitations
It completely relies on the browser extension to implement advanced scraping functions. If the browser is not installed or the extension is not enabled, it cannot handle complex single-page applications or interactive web pages.
The installation and initial configuration are relatively cumbersome. Users need to install the browser, browser extension, Node.js or Python environment, and the MCP server package separately, and ensure that the port configurations are consistent.
The performance is limited by the browser extension and the local network. The scraping speed may not be as high as that of professional, distributed cloud crawlers, and it is not suitable for large-scale, high-concurrency data collection tasks.

How to Use

Install the Browser and Extension
Make sure you have installed the Chrome or Edge browser. Then, install the LionScraper extension in the browser's extension store.
Install the Runtime Environment and MCP Package
Install Node.js 18+ or Python 3.10+. Then, install the LionScraper package via npm or pip. After installation, your system will have the `lionscraper` and `lionscraper-mcp` commands.
Start the Daemon
Run the `lionscraper daemon` command in the terminal to start the background service. This service will listen on a port (default is 13808) for communication with the browser extension and AI application.
Configure the Browser Extension Port
Open the settings or options page of the LionScraper extension, find the 'Bridge Port' setting, and modify its value to the same port as the MCP server (e.g., 13808). If necessary, click 'Reconnect' or restart the browser.
Configure MCP in the AI Application
In the MCP configuration file of the AI application you are using (such as Cursor), add a new server named `lionscraper` and set `command` to `lionscraper-mcp`.
Start Using
After completing the above configurations, in the conversation interface of your AI application, you can directly use natural language to instruct the AI to perform web data collection. For example: 'Connect to LionScraper and then scrape all photo URLs on my current browser page.'

Usage Examples

Collect Competitor's Product Information
Market researchers want to quickly understand all products and their prices on a competitor's website for market analysis. By giving instructions in the AI application, they can scrape the product list with one click.
Organize Industry News Summaries
Content editors need to collect today's headline news titles and links from several different technology news websites to create a daily briefing.
Extract Information from the Contact Page
A salesperson wants to find the company's phone number and email from the 'Contact Us' page of a potential client to make contact.
Backup Blog Article Content
You want to save a long article on your own blog locally, including its body content.

Frequently Asked Questions

Why can I see the MCP tool in the AI application, but the scraping still fails?
Do I need to install Node.js or Python?
How to change the port of LionScraper?
Can LionScraper work without using the browser extension?
Can I run LionScraper on multiple computers?

Related Resources

Official Website
The official website of the LionScraper project, where you can get the latest information and updates.
npm Package
The Node.js version of the LionScraper package, which can be installed via npm.
PyPI Project
The Python version of the LionScraper project, which can be installed via pip.
GitHub Repository
The source code repository on GitHub, containing detailed documentation for the Node.js and Python packages.
Chrome Extension Store
Install the LionScraper extension in the Chrome browser.
Microsoft Edge Extension Store
Install the LionScraper extension in the Edge browser.

Installation

Copy the following command to your Client for configuration
{
  "mcpServers": {
    "lionscraper": {
      "command": "lionscraper-mcp"
    }
  }
}

{
  "mcpServers": {
    "lionscraper": {
      "command": "lionscraper-mcp",
      "env": {
        "PORT": "13808",
        "TIMEOUT": "120000",
        "LANG": "en-US",
        "TOKEN": "",
        "DAEMON": ""
      }
    }
  }
}

{
  "mcpServers": {
    "lionscraper": {
      "command": "npx",
      "args": ["-y", "lionscraper", "lionscraper-mcp"]
    }
  }
}

{
  "mcpServers": {
    "lionscraper": {
      "command": "npx",
      "args": ["-y", "lionscraper", "lionscraper-mcp"],
      "env": {
        "PORT": "13808",
        "TIMEOUT": "120000",
        "LANG": "en-US",
        "TOKEN": "",
        "DAEMON": ""
      }
    }
  }
}

{
  "mcpServers": {
    "lionscraper": {
      "command": "python",
      "args": ["-m", "lionscraper"]
    }
  }
}
Note: Your key is sensitive information, do not share it with anyone.

Alternatives

P
Paperbanana
Python
8.9K
5 points
F
Finlab Ai
FinLab AI is a quantitative financial analysis platform that helps users discover excess returns (alpha) in investment strategies through AI technology. It provides a rich dataset, backtesting framework, and strategy examples, supporting automated installation and integration into mainstream AI programming assistants.
8.8K
4 points
A
Apify MCP Server
The Apify MCP Server is a tool based on the Model Context Protocol (MCP) that allows AI assistants to extract data from websites such as social media, search engines, and e-commerce through thousands of ready-to-use crawlers, scrapers, and automation tools (Apify Actors). It supports OAuth and Skyfire proxy payment and can be integrated into MCP clients such as Claude and VS Code through HTTPS endpoints or local stdio.
TypeScript
8.9K
5 points
P
Praisonai
PraisonAI is a production-ready multi-AI agent framework with self-reflection capabilities, designed to create AI agents to automate the solution of various problems from simple tasks to complex challenges. It simplifies the construction and management of multi-agent LLM systems by integrating PraisonAI agents, AG2, and CrewAI into a low-code solution, emphasizing simplicity, customization, and effective human-machine collaboration.
Python
16.7K
5 points
M
Maverick MCP
MaverickMCP is a personal stock analysis server based on FastMCP 2.0, providing professional level financial data analysis, technical indicator calculation, and investment portfolio optimization tools for MCP clients such as Claude Desktop. It comes pre-set with 520 S&P 500 stock data, supports multiple technical analysis strategies and parallel processing, and can run locally without complex authentication.
Python
11.9K
4 points
K
Klavis
Klavis AI is an open-source project that provides a simple and easy-to-use MCP (Model Context Protocol) service on Slack, Discord, and Web platforms. It includes various functions such as report generation, YouTube tools, and document conversion, supporting non-technical users and developers to use AI workflows.
TypeScript
34.9K
5 points
S
Scrapling
Scrapling is an adaptive web scraping library that can automatically learn website changes and re - locate elements. It supports multiple scraping methods and AI integration, providing high - performance parsing and a developer - friendly experience.
Python
24.9K
5 points
A
Apple Health MCP
An MCP server for querying Apple Health data via SQL, implemented based on DuckDB for efficient analysis, supporting natural language queries and automatic report generation.
TypeScript
13.2K
4.5 points
M
Markdownify MCP
Markdownify is a multi-functional file conversion service that supports converting multiple formats such as PDFs, images, audio, and web page content into Markdown format.
TypeScript
38.1K
5 points
D
Duckduckgo MCP Server
Certified
The DuckDuckGo Search MCP Server provides web search and content scraping services for LLMs such as Claude.
Python
80.3K
4.3 points
G
Gitlab MCP Server
Certified
The GitLab MCP server is a project based on the Model Context Protocol that provides a comprehensive toolset for interacting with GitLab accounts, including code review, merge request management, CI/CD configuration, and other functions.
TypeScript
28.5K
4.3 points
N
Notion Api MCP
Certified
A Python-based MCP Server that provides advanced to-do list management and content organization functions through the Notion API, enabling seamless integration between AI models and Notion.
Python
23.8K
4.5 points
F
Figma Context MCP
Framelink Figma MCP Server is a server that provides access to Figma design data for AI programming tools (such as Cursor). By simplifying the Figma API response, it helps AI more accurately achieve one - click conversion from design to code.
TypeScript
69.6K
4.5 points
U
Unity
Certified
UnityMCP is a Unity editor plugin that implements the Model Context Protocol (MCP), providing seamless integration between Unity and AI assistants, including real - time state monitoring, remote command execution, and log functions.
C#
37.4K
5 points
G
Gmail MCP Server
A Gmail automatic authentication MCP server designed for Claude Desktop, supporting Gmail management through natural language interaction, including complete functions such as sending emails, label management, and batch operations.
TypeScript
24.0K
4.5 points
M
Minimax MCP Server
The MiniMax Model Context Protocol (MCP) is an official server that supports interaction with powerful text-to-speech, video/image generation APIs, and is suitable for various client tools such as Claude Desktop and Cursor.
Python
56.4K
4.8 points
AIBase
Zhiqi Future, Your AI Solution Think Tank
© 2026AIBase