| Episode | Status |
|---|---|
| Episode | Status |
|---|---|
Marek Kozlowski, Head of the AI Lab at Poland's National Information Processing Institute, discusses project PLLuM (Polish Large Language Models). PSA for AI builders: Interested in alignment, governa...
Marek Kozlowski, Head of AI Lab at Poland's National Information Processing Institute, discusses Project PLLuM (Polish Large Language Models), a government-funded initiative to create sovereign AI. He explains how smaller, locally-adapted models can compete with frontier models for specific languages and domains while maintaining control, privacy, and cost advantages. The conversation covers technical strategies like language adaptation on base models, challenges from EU regulations limiting data access, and why the future of enterprise AI may favor specialized on-premise models over massive cloud-based systems.
Kozlowski explains the core thesis of localized LMs: creating models 10x smaller than frontier models but with equivalent quality in specific languages and domains. He discusses how 90% of training data is English/Chinese, leading to cultural and linguistic gaps in other languages, and why models trained primarily on English produce non-native sounding text in Polish despite being technically correct.
Discussion of how EU AI Act and copyright regulations create much harder constraints than US companies face, forcing European teams to rely on curated datasets rather than massive web scrapes. Kozlowski explains the data curation process including deduplication and filtering that can reduce datasets by 50-80%, and why they don't have the trillion tokens needed for training from scratch.
Kozlowski details PLLuM's competitive advantage: dozens to hundreds of human annotators creating organic instructions and preferences rather than relying on synthetic data. He explains why synthetic instructions from other LLMs can degrade model quality if linguistically poor, and how they've built internal tools rather than using commercial annotation platforms.
Technical explanation of how PLLuM uses language adaptation rather than training from scratch. They take LLAMA or Mistral base models and continue pre-training on Polish corpora, creating new base models that retain some English capabilities while gaining Polish fluency. Discusses the forgetting problem in cascade learning and why 1 trillion tokens are needed for stable 8B parameter models.
Critical finding: Anthropic Claude and OpenAI GPT models are showing declining performance in Polish language and cultural competency across successive releases. As companies optimize for high-value use cases like coding, they're deprioritizing niche language quality, creating integration risk for businesses that depend on these models.
Kozlowski explains domain adaptation (continued pre-training on company-specific data) and its requirements. Working with major European bank PKO BP, they found you need 10B+ tokens after deduplication (30-40B before) to make domain adaptation worthwhile - limiting this approach to only the largest companies with sufficient proprietary data.
Strategic argument for why AI's future favors small, specialized models over massive general-purpose ones. Businesses typically need 10-20 use cases, not 1000+ tasks. With 1000+ training examples per task, supervised fine-tuning of smaller models achieves same quality as few-shot prompting of large cloud models, with better cost, control, and regulatory compliance.
Unlike US companies developing detailed AI constitutions, Poland hasn't needed explicit value articulation yet. Current focus is on ethical behavior and data usage regulations rather than behavioral constraints. Kozlowski notes they're a few years behind US/China in development, with regulations focused on what data can be used rather than how models should behave.
Poland's government is prioritizing GPU infrastructure (AI factories) over competing with US salaries for top talent. Kozlowski explains the challenge of competing when US AI engineers earn 'NFL quarterback' salaries, and why public sector projects struggle to define monetizable objectives that would justify such compensation.
Kozlowski observes that frontier model improvements are becoming more incremental (plateau effect) rather than the shocking leaps of 2022-2024. He argues the industry is entering a verification phase where businesses will evaluate real costs and benefits, likely concluding they don't need massive general-purpose models for their specific use cases.
Discussion of the fundamental tension between distributed cooperation (ideal for spreading wealth and knowledge) and centralization (necessary for fast outcomes). Kozlowski uses Silicon Valley as example: while distributed startups across US would be economically better, concentration enables rapid innovation. Questions whether US-China rivalry allows for cooperation.
Sovereign AI in Poland: Language Adaptation, Local Control & Cost Advantages with Marek Kozlowski
Ask me anything about this podcast episode...
Try asking: