Allow loading remote contents and showing images to get the best out of this email.FAUN.dev's AI/ML Weekly Newsletter
 
🔗 View in your browser   |  ✍️ Publish on FAUN.dev   |  🦄 Become a sponsor
 
Allow loading remote contents and showing images to get the best out of this email.
Kala
 
#ArtificialIntelligence #MachineLearning #MLOps
 
 
📝 A Few Words
 
 
AI voice is catching up fast !

Qwen3-TTS the open source text-to-speech model from Qwen.ai is out and it's a big deal!

3 details in the latest release really change the conversation around AI voice:

  • 1st: ~97 ms end-to-end streaming latency => That's no longer "fast TTS", it's interactive.
  • 2nd: voice cloning from ~3 seconds of reference audio => That dramatically lowers the cost of personalization.
  • 3rd: 10-language support with cross-lingual cloning. This is subtle but important.

Put together, this starts to look like a real shift.

AI Voice used to be a premium, closed, API-only feature but it's becoming infrastructure. And once voice becomes infrastructure, a lot of products that currently default to text will quietly start talking.

The fact that this is open source is huge as it puts pressure on closed providers like ElevenLabs and accelerates experimentation.

We're probably going to see a wave of real-time voice agents that don't feel like demos anymore!

Have a great week!
Aymen
 
 
🔍 Inside this Issue
 
 
Agents are crashing the ops party while security draws the lines: AWS drops Frontier Agents, MCP tightens auth, and Palantir shows what real guardrails look like. Meanwhile Modal keeps 20k GPUs alive, OpenClaw eats a security wake-up, and the cloud GPU market gets sharper edges; details inside.

🤖 AWS Frontier Agents: Kiro, DevOps Agent, and Security Agent
🔐 Is that allowed? Authentication and authorization in Model Context Protocol
🩺 Keeping 20,000 GPUs healthy
⚠️ OpenClaw - Former Moltbot, Former Clawdbot - Went Viral Overnight. Then Security Reality Hit.
🛡️ Securing Agents in Production (Agentic Runtime, #1)
☁️ Top 8 Cloud GPU Providers for AI and Machine Learning

Ship smarter this week.

Thanks for reading!
FAUN.dev() Team
 
 
ℹ️ News, Updates & Announcements
 
faun.dev faun.dev
 
OpenClaw - Former Moltbot, Former Clawdbot - Went Viral Overnight. Then Security Reality Hit.
 
 
OpenClaw just dropped the corporate mask. New name, new direction. Support now spans Twitch, Google Chat, plus toys like KIMI K2.5 and MiMo-V2-Flash. Still all in on self-hosting, maybe even more so.

Version 2026.1.29 patched a nasty auth bypass flaw (CVE-2026-25253). Gateway connection checks are tighter now. Better late than never.
 
 
👉 Enjoyed this?Read more news on FAUN.dev/news
 
Stories From The Chief I/O
 
thechief.io thechief.io
 
Top 8 Cloud GPU Providers for AI and Machine Learning
 
 
A fresh dive into eight top cloud GPU platforms shows how the game’s changing, fast. Giants like AWS, GCP, and Azure now rub shoulders with scrappier upstarts like TensorDock and RunPod. What’s on the menu? A100/H100 GPUs, bare-metal access, tighter MLOps hooks, and pricing sliced thinner than ever.
 
 
 
⭐ Sponsors
 
faun.dev faun.dev
 
🚀 Take Your Coding Experience to the Next Level !
 
 
Uncover how to level-up your GitHub Copilot and VS Code experience from an autocomplete assistant to an intelligent, agentic teammate that can navigate your codebase, execute tasks, reason across files and even manage your GitHub projects.

In Building with GitHub Copilot course, you're not just learning how to use GitHub Copilot. You're exploring a shift in how we write, reason about, and collaborate on code.

Get your copy today!
 
 
👉 Spread the word and help developers find you by promoting your projects on FAUN. Get in touch for more information.
 
🔗 Stories, Tutorials & Articles
 
stackoverflow.blog stackoverflow.blog
 
Is that allowed? Authentication and authorization in Model Context Protocol
 
 
The Model Context Protocol (MCP) 2025-11-25 spec tightens up remote agent auth. It leans into OAuth 2.1 Authorization Code grants, PKCE required, step-up auth backed. No token passthrough allowed.

What’s new: experimental extensions for client credentials and client ID metadata. These smooth out agent registration and grease the wheels for server-to-server auth.
 
 
modal.com modal.com
 
Keeping 20,000 GPUs healthy
 
 
Modal unpacked how it keeps a 20,000+ GPU fleet sane across AWS, GCP, Azure, and OCI. Think autoscaling, yes, but with some serious moves behind the curtain.

They're running instance benchmarking, enforcing machine image consistency, running boot-time checks, and tracking GPU health both passively and actively. Sick GPUs get quarantined. The whole thing’s wired up with image testing and auto-failover.
 
 
blog.palantir.com blog.palantir.com
 
Securing Agents in Production (Agentic Runtime, #1)
 
 
Palantir's AIP Agentic Runtime isn't just another agent platform, it's a control plane with teeth. Think tight policy enforcement, ephemeral autoscaling with Kubernetes (Rubix), and memory stitched in from the jump via Ontology.

Tool usage? Traced and locked down with provenance-based security. Every LLM prompt, function call, and data tweak? Logged, versioned, and fully observable.
 
 
newsletter.simpleaws.dev newsletter.simpleaws.dev
 
AWS Frontier Agents: Kiro, DevOps Agent, and Security Agent
 
 
“Frontier Agents” drop straight into incident workflows. They kick off investigations on their own, whether triggered by alarms or a human hand, pulling together logs, metrics, and deployment context fast.

Findings show up where they’re needed: Slack threads, tickets, operator dashboards. No shell commands. No vague prompts. Just answers.
 
 

👉 Got something to share? Create your FAUN Page and start publishing your blog posts, tools, and updates. Grow your audience, and get discovered by the developer community.

 
⭐ Supporters
 
bytevibe.co bytevibe.co
 
Take a break!
 
 
Take a break, and get a coffee! Warm your soul with a nice mug perfectly sized black ceramic mug.
 
 
👉 Spread the word and help developers find you by promoting your projects on FAUN. Get in touch for more information.
 
⚙️ Tools, Apps & Software
 
github.com github.com
 
different-ai/openwork
 
 
An open-source alternative to Claude Cowork, powered by OpenCode
 
 
github.com github.com
 
affaan-m/everything-claude-code
 
 
Complete Claude Code configuration collection - agents, skills, hooks, commands, rules, MCPs. Battle-tested configs from an Anthropic hackathon winner.
 
 
github.com github.com
 
openclaw/openclaw
 
 
Your own personal AI assistant. Any OS. Any Platform. The lobster way.
 
 
github.com github.com
 
tomdyson/toktab.com
 
 
Current pricing data for 2000+ AI models
 
 
github.com github.com
 
cloudflare/moltworker
 
 
Run OpenClaw, (formerly Moltbot, formerly Clawdbot) on Cloudflare Workers
 
 

👉 Spread the word and help developers find and follow your Open Source project by promoting it on FAUN. Get in touch for more information.

 
🤔 Did you know?
 
 
Did you know vLLM leverages its PagedAttention mechanism to rethink how attention key-value (KV) caches are managed during LLM serving, slicing each sequence’s KV cache into small, fixed-size blocks and mapping them via a page-table-like structure rather than reserving huge contiguous slabs? This block-based approach nearly eliminates internal and external memory waste and lets vLLM schedule and batch requests dynamically, which translates into 2–4× higher throughput compared with conventional serving systems like FasterTransformer or Orca at similar latency. The key insight is that the performance gain stems from memory management and scheduling improvements - tighter GPU memory utilization and continuous batching - rather than changing the underlying model math.
 
 
🤖 Once, SenseiOne Said
 
 
"We demand reproducibility from a stochastic process; MLOps is the unglamorous work that makes that demand enforceable. If you can't roll back a model with its data, your system is an experiment, not a service."
- SenseiOne
 

(*) SenseiOne is FAUN.dev’s work-in-progress AI agent

 
⚡Growth Notes
 
 
The people who quietly pull ahead in ML are the ones who rebuild key papers’ results from scratch on their own stack, not just re-run public repos, because that habit exposes you to the messy failure modes everyone hand-waves past (bad seeds, data leakage, flaky evaluations, broken baselines). Do this regularly, and you stop being the person who "tries models" and start becoming the one who decides which results a team can safely bet months of engineering effort on.
 
Each week, we share a practical move to grow faster and work smarter
 
😂 Meme of the week
 
 
 
 
❤️ Thanks for reading
 
 
👋 Keep in touch and follow us on social media:
- 💼LinkedIn
- 📝Medium
- 🐦Twitter
- 👥Facebook
- 📰Reddit
- 📸Instagram

👌 Was this newsletter helpful?
We'd really appreciate it if you could forward it to your friends!

🙏 Never miss an issue!
To receive our future emails in your inbox, don't forget to add community@faun.dev to your contacts.

🤩 Want to sponsor our newsletter?
Reach out to us at sponsors@faun.dev and we'll get back to you as soon as possible.
 

Kala #514: Keeping 20,000 GPUs Healthy
Legend: ✅ = Editor's Choice / ♻️ = Old but Gold / ⭐ = Promoted / 🔰 = Beginner Friendly

You received this email because you are subscribed to FAUN.dev.
We (🐾) help developers (👣) learn and grow by keeping them up with what matters.

You can manage your subscription options here (recommended) or use the old way here (legacy). If you have any problem, read this or reply to this email.