Track MCP LogoTrack MCP
Track MCP LogoTrack MCP

The world's largest repository of Model Context Protocol servers. Discover, explore, and submit MCP tools.

Product

  • Categories
  • Top MCP
  • New & Updated

Company

  • About

Legal

  • Privacy Policy
  • Terms of Service
  • Cookie Policy

© 2025 TrackMCP. All rights reserved.

Built with ❤️ by Krishna Goyal

    Lara Mcp

    77 stars
    TypeScript
    Updated Nov 4, 2025

    Documentation

    Lara Translate MCP Server

    A Model Context Protocol (MCP) Server for Lara Translate API, enabling powerful translation capabilities with support for language detection, context-aware translations and translation memories.

    License

    Docker Pulls

    npm downloads

    📚 Table of Contents

    • 📖 Introduction
    • 🛠 Available Tools
    • 🚀 Getting Started
    • 📋 HTTP Server
    • 🔌 STDIO Server
    • 🧪 Verify Installation
    • 💻 Popular Clients that supports MCPs
    • 🆘 Support

    📖 Introduction

    What is MCP?

    Model Context Protocol (MCP) is an open standardized communication protocol that enables AI applications to connect with external tools, data sources, and services. Think of MCP like a USB-C port for AI applications - just as USB-C provides a standardized way to connect devices to various peripherals, MCP provides a standardized way to connect AI models to different data sources and tools.

    Lara Translate MCP Server enables AI applications to access Lara Translate's powerful translation capabilities through this standardized protocol.

    More info about Model Context Protocol on: https://modelcontextprotocol.io/

    How Lara Translate MCP Works

    Lara Translate MCP Server implements the Model Context Protocol to provide seamless translation capabilities to AI applications. The integration follows this flow:

    1. Connection Establishment: When an MCP-compatible AI application starts, it connects to configured MCP servers, including the Lara Translate MCP Server

    2. Tool & Resource Discovery: The AI application discovers available translation tools and resources provided by the Lara Translate MCP Server

    3. Request Processing: When translation needs are identified:

    • The AI application formats a structured request with text to translate, language pairs, and optional context
    • The MCP server validates the request and transforms it into Lara Translate API calls
    • The request is securely sent to Lara Translate's API using your credentials

    4. Translation & Response: Lara Translate processes the translation using advanced AI models

    5. Result Integration: The translation results are returned to the AI application, which can then incorporate them into its response

    This integration architecture allows AI applications to access professional-grade translations without implementing the API directly, while maintaining the security of your API credentials and offering flexibility to adjust translation parameters through natural language instructions.

    Why to use Lara inside an LLM

    Integrating Lara with LLMs creates a powerful synergy that significantly enhances translation quality for non-English languages.

    Why General LLMs Fall Short in Translation

    While large language models possess broad linguistic capabilities, they often lack the specialized expertise and up-to-date terminology required for accurate translations in specific domains and languages.

    Lara’s Domain-Specific Advantage

    Lara overcomes this limitation by leveraging Translation Language Models (T-LMs) trained on billions of professionally translated segments. These models provide domain-specific machine translation that captures cultural nuances and industry terminology that generic LLMs may miss. The result: translations that are contextually accurate and sound natural to native speakers.

    Designed for Non-English Strength

    Lara has a strong focus on non-English languages, addressing the performance gap found in models such as GPT-4. The dominance of English in datasets such as Common Crawl and Wikipedia results in lower quality output in other languages. Lara helps close this gap by providing higher quality understanding, generation, and restructuring in a multilingual context.

    Faster, Smarter Multilingual Performance

    By offloading complex translation tasks to specialized T-LMs, Lara reduces computational overhead and minimizes latency—a common issue for LLMs handling non-English input. Its architecture processes translations in parallel with the LLM, enabling for real-time, high-quality output without compromising speed or efficiency.

    Cost-Efficient Translation at Scale

    Lara also lowers the cost of using models like GPT-4 in non-English workflows. Since tokenization (and pricing) is optimized for English, using Lara allows translation to take place before hitting the LLM, meaning that only the translated English content is processed. This improves cost efficiency and supports competitive scalability for global enterprises.

    🛠 Available Tools

    Translation Tools

    translate - Translate text between languages

    Inputs:

    • text (array): An array of text blocks to translate, each with:
    • text (string): The text content
    • translatable (boolean): Whether this block should be translated
    • source (optional string): Source language code (e.g., 'en-EN')
    • target (string): Target language code (e.g., 'it-IT')
    • context (optional string): Additional context to improve translation quality
    • instructions (optional string[]): Instructions to adjust translation behavior
    • source_hint (optional string): Guidance for language detection

    Returns: Translated text blocks maintaining the original structure

    Translation Memories Tools

    list_memories - List saved translation memories

    Returns: Array of memories and their details

    create_memory - Create a new translation memory

    Inputs:

    • name (string): Name of the new memory
    • external_id (optional string): ID of the memory to import from MyMemory (e.g., 'ext_my_[MyMemory ID]')

    Returns: Created memory data

    update_memory - Update translation memory name

    Inputs:

    • id (string): ID of the memory to update
    • name (string): The new name for the memory

    Returns: Updated memory data

    delete_memory - Delete a translation memory

    Inputs:

    • id (string): ID of the memory to delete

    Returns: Deleted memory data

    add_translation - Add a translation unit to memory

    Inputs:

    • id (string | string[]): ID or IDs of memories where to add the translation unit
    • source (string): Source language code
    • target (string): Target language code
    • sentence (string): The source sentence
    • translation (string): The translated sentence
    • tuid (optional string): Translation Unit unique identifier
    • sentence_before (optional string): Context sentence before
    • sentence_after (optional string): Context sentence after

    Returns: Added translation details

    delete_translation - Delete a translation unit from memory

    Inputs:

    • id (string): ID of the memory
    • source (string): Source language code
    • target (string): Target language code
    • sentence (string): The source sentence
    • translation (string): The translated sentence
    • tuid (optional string): Translation Unit unique identifier
    • sentence_before (optional string): Context sentence before
    • sentence_after (optional string): Context sentence after

    Returns: Removed translation details

    import_tmx - Import a TMX file into a memory

    Inputs:

    • id (string): ID of the memory to update
    • tmx_content (string): The content of the tmx file to upload
    • gzip (boolean): Indicates if the file is compressed (.gz)

    Returns: Import details

    check_import_status - Checks the status of a TMX file import

    Inputs:

    • id (string): The ID of the import job

    Returns: Import details

    🚀 Getting Started

    Lara supports both the STDIO and streamable HTTP protocols. For a hassle-free setup, we recommend using the HTTP protocol. If you prefer to use STDIO, it must be installed locally on your machine.

    You'll find setup instructions for both protocols in the sections below.

    HTTP Server 🌐

    ❌ Clients NOT supporting url configuration (e.g., Claude, OpenAI)

    This installation guide is intended for clients that do NOT support the url-based configuration. This option requires Node.js to be installed on your system.

    If you're unsure how to configure an MCP with your client, please refer to your MCP client's official documentation.

    ---

    1. Open your client's MCP configuration JSON file with a text editor, then copy and paste the following snippet:

    json
    {
      "mcpServers": {
        "lara": {
          "command": "npx",
          "args": [
            "mcp-remote",
            "https://mcp.laratranslate.com/v1",
            "--header",
            "x-lara-access-key-id: ${X_LARA_ACCESS_KEY_ID}",
            "--header",
            "x-lara-access-key-secret: ${X_LARA_ACCESS_KEY_SECRET}"
          ],
          "env": {
            "X_LARA_ACCESS_KEY_ID": "",
            "X_LARA_ACCESS_KEY_SECRET": ""
          }
        }
      }
    }

    2. Replace ` and ` with your Lara Translate API credentials. Refer to the Official Documentation for details.

    3. Restart your MCP client.

    ✅ Clients supporting url configuration (e.g., Cursor, Continue)

    This installation guide is intended for clients that support the url-based configuration. These clients can connect to Lara through a remote HTTP endpoint by specifying a simple configuration object.

    Some examples of supported clients include Cursor, Continue, OpenDevin, and Aider.

    If you're unsure how to configure an MCP with your client, please refer to your MCP client's official documentation.

    ---

    1. Open your client's MCP configuration JSON file with a text editor, then copy and paste the following snippet:

    json
    {
      "mcpServers": {
        "lara": {
          "url": "https://mcp.laratranslate.com/v1",
          "headers": {
            "x-lara-access-key-id": "",
            "x-lara-access-key-secret": ""
          }
        }
      }
    }

    2. Replace ` and ` with your Lara Translate API credentials. Refer to the Official Documentation for details.

    3. Restart your MCP client.

    ---

    STDIO Server 🖥️

    Using NPX

    This option requires Node.js to be installed on your system.

    1. Add the following to your MCP configuration file:

    json
    {
      "mcpServers": {
        "lara-translate": {
          "command": "npx",
          "args": ["-y", "@translated/lara-mcp@latest"],
          "env": {
            "LARA_ACCESS_KEY_ID": "",
            "LARA_ACCESS_KEY_SECRET": ""
          }
        }
      }
    }

    2. Replace ` and ` with your actual Lara API credentials.

    Using Docker

    This option requires Docker to be installed on your system.

    1. Add the following to your MCP configuration file:

    json
    {
      "mcpServers": {
        "lara-translate": {
          "command": "docker",
          "args": [
            "run",
            "-i",
            "--rm",
            "-e",
            "LARA_ACCESS_KEY_ID",
            "-e",
            "LARA_ACCESS_KEY_SECRET",
            "translatednet/lara-mcp:latest"
          ],
          "env": {
            "LARA_ACCESS_KEY_ID": "",
            "LARA_ACCESS_KEY_SECRET": ""
          }
        }
      }
    }

    2. Replace ` and ` with your actual Lara API credentials.

    Building from Source

    Using Node.js

    1. Clone the repository:

    bash
    git clone https://github.com/translated/lara-mcp.git
    cd lara-mcp

    2. Install dependencies and build:

    bash
    # Install dependencies
    pnpm install
    
    # Build
    pnpm run build

    3. Add the following to your MCP configuration file:

    json
    {
      "mcpServers": {
        "lara-translate": {
          "command": "node",
          "args": ["/dist/index.js"],
          "env": {
            "LARA_ACCESS_KEY_ID": "",
            "LARA_ACCESS_KEY_SECRET": ""
          }
        }
      }
    }

    4. Replace:

    • `` with the absolute path to your project folder
    • ` and ` with your actual Lara API credentials.

    Building a Docker Image

    1. Clone the repository:

    bash
    git clone https://github.com/translated/lara-mcp.git
    cd lara-mcp

    2. Build the Docker image:

    bash
    docker build -t lara-mcp .

    3. Add the following to your MCP configuration file:

    json
    {
      "mcpServers": {
        "lara-translate": {
          "command": "docker",
          "args": [
            "run",
            "-i",
            "--rm",
            "-e",
            "LARA_ACCESS_KEY_ID",
            "-e",
            "LARA_ACCESS_KEY_SECRET",
            "lara-mcp"
          ],
          "env": {
            "LARA_ACCESS_KEY_ID": "",
            "LARA_ACCESS_KEY_SECRET": ""
          }
        }
      }
    }

    4. Replace ` and ` with your actual credentials.

    🧪 Verify Installation

    After restarting your MCP client, you should see Lara Translate MCP in the list of available MCPs.

    The method for viewing installed MCPs varies by client. Please consult your MCP client's documentation.

    To verify that Lara Translate MCP is working correctly, try translating with a simple prompt:

    text
    Translate with Lara "Hello world" to Spanish

    Your MCP client will begin generating a response. If Lara Translate MCP is properly installed and configured, your client will either request approval for the action or display a notification that Lara Translate is being used.

    💻 Popular Clients that supports MCPs

    For a complete list of MCP clients and their feature support, visit the official MCP clients page.

    ClientDescription
    Claude DesktopDesktop application for Claude AI
    AixplainProduction-ready AI Agents
    CursorAI-first code editor
    Cline for VS CodeVS Code extension for AI assistance
    GitHub Copilot MCPVS Code extension for GitHub Copilot MCP integration
    WindsurfAI-powered code editor and development environment

    🆘 Support

    • For issues with Lara Translate API: Visit Lara Translate API and Integrations Support
    • For issues with this MCP Server: Open an issue on GitHub

    Similar MCP

    Based on tags & features

    • OP

      Openai Gpt Image Mcp

      TypeScript·
      75
    • MC

      Mcgravity

      TypeScript·
      71
    • PL

      Pluggedin Mcp Proxy

      TypeScript·
      97
    • MC

      Mcp Open Library

      TypeScript·
      42

    Trending MCP

    Most active this week

    • PL

      Playwright Mcp

      TypeScript·
      22.1k
    • SE

      Serena

      Python·
      14.5k
    • MC

      Mcp Playwright

      TypeScript·
      4.9k
    • MC

      Mcp Server Cloudflare

      TypeScript·
      3.0k
    View All MCP Servers

    Similar MCP

    Based on tags & features

    • OP

      Openai Gpt Image Mcp

      TypeScript·
      75
    • MC

      Mcgravity

      TypeScript·
      71
    • PL

      Pluggedin Mcp Proxy

      TypeScript·
      97
    • MC

      Mcp Open Library

      TypeScript·
      42

    Trending MCP

    Most active this week

    • PL

      Playwright Mcp

      TypeScript·
      22.1k
    • SE

      Serena

      Python·
      14.5k
    • MC

      Mcp Playwright

      TypeScript·
      4.9k
    • MC

      Mcp Server Cloudflare

      TypeScript·
      3.0k