Skip to content

Markdown Generation for LLM Integration #2098

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
andyl opened this issue Mar 18, 2025 · 2 comments
Closed

Markdown Generation for LLM Integration #2098

andyl opened this issue Mar 18, 2025 · 2 comments

Comments

@andyl
Copy link

andyl commented Mar 18, 2025

LLMs perform well on languages like Python and Javascript, where there is a great deal of training data. LLMs perform less well on Elixir because there is a smaller volume of training data. It can take months between LLM releases, and sometimes new and rapidly changing Elixir libraries like ASH are not properly incorporated in the training data.

Feature request: for LLM integration, it would be nice if ExDoc generated markdown-formatted variants (*.md) for each page.

These markdown files can be ingested at 'run time' by LLMs and AI Development tools.

The potential benefit for Elixir developers is to make HEX packages first-class LLM citizens, even for very new and rapidly changing libraries.

Comparables and Reference Information

Overview: LLMs integration with LLMS.TXT protocol

  1. Claude (Anthropic)

    • Description: Anthropic’s Claude, a conversational AI model, supports llms.txt files by allowing users to upload or paste their contents directly into the chat interface. This enables Claude to utilize the concise, LLM-friendly context provided by llms.txt for more accurate responses about a website or project.
    • How It Uses llms.txt: Users can manually provide the llms.txt file content (e.g., via copy-paste or file upload) to give Claude up-to-date, structured context, especially useful for technical documentation or APIs. Anthropic has implemented basic support, as seen with their own /llms.txt file.
    • Purpose: Enhances Claude’s ability to reason over specific site content without parsing raw HTML or dealing with context window limitations.
  2. ChatGPT (OpenAI)

    • Description: ChatGPT can process llms.txt files if users input the file content or a URL pointing to it (where supported by browsing capabilities in certain versions).
    • How It Uses llms.txt: By pasting the contents of an llms.txt file into the prompt, ChatGPT can use the markdown-structured data to answer questions or perform tasks related to the site or project described. It doesn’t natively crawl for llms.txt, but it leverages the format when provided.
    • Purpose: Allows users to give ChatGPT a distilled version of a website’s key information, improving response relevance for specific queries.

Overview: AI development tool integration with LLMS.TXT protocol

  1. Cursor

    • Description: An AI-first code editor built on VS Code, designed for efficient coding with LLM integration.
    • How It Uses llms.txt: Supports adding llms.txt files via the @Docs feature. Users can manually input or link to an llms.txt file, indexing it as context for chats and code generation, enhancing project-specific assistance.
    • Purpose: Improves contextual coding by referencing structured, LLM-friendly documentation without parsing raw HTML.
  2. Windsurf

    • Description: An agentic IDE by Codeium, combining copilot and agent features for multi-file editing and contextual awareness.
    • How It Uses llms.txt: While not natively fetching llms.txt, users can paste its contents into the Cascade agent or chat interface, leveraging the markdown format for precise, project-specific responses.
    • Purpose: Enhances agent-driven coding by utilizing concise, structured data from llms.txt for better codebase understanding.
  3. Aider

    • Description: A terminal-based AI pair programming tool that edits code in local Git repositories using LLMs.
    • How It Uses llms.txt: Users can manually provide llms.txt content via natural language prompts, allowing Aider to incorporate project documentation or metadata into its multi-file editing process.
    • Purpose: Boosts efficiency in refactoring and feature implementation by grounding LLM actions in llms.txt-provided context.
@josevalim
Copy link
Member

Please see #1976 and #1992 and associated discussions.

@josevalim
Copy link
Member

Closing in favor of the on-going PRs, thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

2 participants