Updates
Oct 6, 2025
Highlights: Antler × OpenAI × 913.ai Hackathon in Munich
We hosted our first AI Agents Hackathon in Munich together with Antler and OpenAI - right by Theresienwiese during Oktoberfest. OpenAI sponsored credits, Antler provided the space and supported the ecosystem push, and we brought three enterprise challenges that our customers care about daily.
60+ participants selected from hundreds of applications
Cross-disciplinary talent: doctors, mathematicians, data scientists, engineers—the brightest minds in Munich
One intense day of building real, working solutions
Three winning teams - congratulations:
1st: Tilman Resch, Daniel Lindner, Jay Kalaria
2nd: Yogesh Kawadkar
3rd: Phi Linh Phan, Francisco Kusch Domínguez
As Maveen, our CTO, shared: hackathons aren’t just about code - they’re about choosing the right problems and giving brilliant people clear challenges. When you do that, you get results that push the frontier.
The Three Tracks: Real Enterprise Problems, Real Progress
1) Beyond RAG: Processing Very Long Documents
Enterprise teams live in long-form content—contracts, playbooks, SOPs, policy manuals. Standard RAG often breaks under this weight because the “right” context isn’t fully or correctly retrieved.
Core issues we set out to tackle:
Missing or partial context across large documents
Weak structure awareness (sections, clauses, annexes)
Latency and token constraints in long-context workflows
Emerging directions builders explored:
Hierarchical retrieval and semantic chunking with overlaps to preserve section meaning
Context compression and pruning for long sequences to keep only evidentially relevant content
Multi-pass retrieval with diversity to avoid brittle single-shot fetches
Agentic orchestration to route between clause-level retrieval and section-level comparison
Why it matters: Long-document accuracy isn’t optional in legal and operations. Teams need grounded answers, traceability, and speed—at once.
References for further reading:
Long-document RAG best practices: semantic chunking and hierarchical indexing, context compression and pruning, diversity-aware retrieval, and dynamic context optimization (see representative discussions and papers on topics like evidential context compression, attention-guided pruning, and dynamic context optimization in 2025 research conversations).
2) Format-Preserving Editing in Word With AI
AI can rewrite and improve content fast—but maintaining exact formatting in Word templates, clause labels, styles, numbering, tables, and cross-references is still a sticking point.
Friction we targeted:
Losing template styles or list hierarchies after edits
Breaking references, TOCs, and complex layouts
No transparent “what changed” view for reviewers
Patterns participants tested:
Strict style-bound edits: generating content that adheres to predefined Word styles and schema
Side-by-side diffs and tracked changes to preserve reviewer trust
Programmatic safeguards: protecting headers/footers, tables, and numbering while allowing content edits
Why it matters: In regulated and enterprise environments, format is part of compliance and brand integrity—not just aesthetics.
Reference examples:
Real-world focus on preserving format and tracked changes in Word during AI-assisted editing, with tooling trends emphasizing “keep formatting” polish and robust change extraction.
3) True Redlining for Legal and Enterprise Negotiations
In legal and procurement, you never start from scratch. You work from a version, add changes, and need redlines that are precise and reviewable.
Gaps we addressed:
Clean, clause-aware redlining—beyond raw text diff
Semantically meaningful comparisons (substance over punctuation/spacing)
Playbook-aligned suggestions that reflect organizational positions
What builders explored:
Semantic diffing to highlight material changes while minimizing noise
Plain-language summaries of changes by clause
Playbook-driven proposals and risk flags that fit how legal teams actually negotiate
Why it matters: Redlining is where time is won or lost. Better diffs + better suggestions = faster deals, less risk, and happier teams.
Reference examples:
Best-practice redlining includes AI-assisted clause analysis, collaborative track changes, integration with Word, and playbook alignment.
What We Learned About Building Domain-Specific Agents
Focus beats generality: Agents that are pre-trained and pre-configured for legal and operations outperform generic prompts when stakes and structures are high.
Structure-aware pipelines win: Treat documents as structured artifacts (sections, clauses, exhibits), not just text blobs.
Trust tooling is non-negotiable: Track changes, cite sources, preserve formatting, and summarize deltas. These are table stakes for enterprise adoption.
Reduce friction for operators: The user experience must feel native to existing tools and workflows.
Our commitment: deliver the most sophisticated domain-specific pre-trained agents for legal and enterprise operations - with as little friction as possible for operational teams.
Event Atmosphere: Building Near Theresienwiese During Oktoberfest
There’s something special about building near Theresienwiese in Oktoberfest season. The energy was palpable - teams formed fast, ideas crystallized even faster, and prototypes emerged that we’re excited to keep pushing forward with the community.
Thank You to Our Partners
Antler: the space and for championing the ecosystem and early builders
OpenAI: for credits and support
The Munich community: for showing up and building with ambition and heart
Increase productivity & efficiency
Human-quality work
Scale & adapt quickly
Meet growing workloads and changing priorities instantly—without the cost or delays of hiring.