8000 Best LLMs for use_llm mode in hybrid Marker pipeline? · Issue #680 · datalab-to/marker · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
Best LLMs for use_llm mode in hybrid Marker pipeline? #680
Open
@eithannak29

Description

@eithannak29

Hi 👋

I'm currently using Marker with use_llm=True and a local Ollama setup. I'm experimenting with various models like deepseek-r1:14b, llama3, etc. for high-quality PDF conversion (scientific documents, multi-column, formulas, tables).

I wanted to ask:

  1. From your internal benchmarks, what LLMs perform best in hybrid mode (Marker + LLM)?
  2. Are there specific models (e.g. Gemini, Claude, Deepseek, OpenAI, etc.) that you've found significantly outperform others in table handling, math rendering, or section parsing?
  3. Any examples of LLM-judge or heuristic scores for use_llm=True configurations?

Thanks a lot — I'm using Marker in a research context and want the most accurate setup, even if it's slow 🚀

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions

      0