TechTalks with Manoj cover art

TechTalks with Manoj

TechTalks with Manoj

By: Powered by the Cloud Driven by Code
Listen for free

About this listen

From code to cloud to cognitive services — TechTalks with Manoj explores the cutting edge of software development. Hosted by a veteran architect with 18+ years in .NET, Angular, and cloud platforms like Azure and AWS, this show is your blueprint to building scalable, modern, and AI-driven applications.

manojknewsletter.substack.comManoj Karkera
Politics & Government
Episodes
  • Demystifying gRPC — The Architecture Behind High-Performance Microservices
    Nov 21 2025

    Welcome back to TechTalks with Manoj — the show where we cut through the hype and talk about the real engineering that makes today’s cloud systems fast, reliable, and production-ready.

    Today, we’re diving into something developers love to name-drop but very few truly understand end to end: gRPC.

    You’ve probably heard “gRPC is faster because it’s binary.”Sure — but that’s barely scratching the surface. The real story goes deeper into transport protocols, schema design, flow control, and the kind of resilience you only appreciate once your system starts sweating under real traffic.

    Think of gRPC as the evolution of service-to-service communication. Not just an API framework — but a more disciplined, more efficient contract between microservices. It brings structure where REST gives flexibility, and speed where JSON gives readability. Most importantly, it gives architects the tools to build systems that behave consistently even when everything around them is under pressure.

    In this episode, we’ll unpack:

    * Why HTTP/2 — and eventually HTTP/3 — are the true engines behind gRPC’s performance.

    * How Protocol Buffers enforce strong contracts while keeping payloads incredibly small.

    * The streaming capabilities that turn gRPC into a real-time powerhouse — and the backpressure rules that keep it from collapsing.

    * Why modern Zero Trust architectures lean on mTLS, JWT, and gateways like Envoy to secure gRPC traffic.

    * The underrated superpower: client-side load balancing, retries, and circuit breakers — and how xDS turns all of this into a centrally managed control plane.

    * And yes, how gRPC compares with REST and gRPC-Web, and when you shouldn’t use it.

    By the end of this episode, you’ll see that gRPC isn’t just a “faster API.”It’s a complete architectural philosophy built for systems that need to be efficient, predictable, and scalable from day one.

    So if you’ve ever wondered how high-performance microservices really talk to each other — this one’s for you.

    Let’s get into it. ⚙️

    Thanks for reading! Subscribe for free to receive new posts and support my work.



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit manojknewsletter.substack.com
    Show More Show Less
    13 mins
  • Microsoft Agent Framework Explained — The Backbone of Enterprise-Grade AI
    Nov 15 2025

    Welcome back to TechTalks with Manoj — the place where we skip the buzzwords and dig into the engineering that actually matters.

    Today, we’re diving into one of the most important — but surprisingly overlooked — pillars of Microsoft’s AI strategy: the Microsoft Agent Framework, better known as MAF.

    If you’ve been wondering how enterprises will move from building flashy AI demos to running reliable, governed, production-grade AI systems… this is the missing piece.

    Think of MAF as Microsoft’s blueprint for bringing order to the wild west of agents — standardizing how they’re built, orchestrated, monitored, and trusted across the enterprise.

    This isn’t just another SDK drop. It’s Microsoft’s attempt to unify everything: tooling, governance, observability, security, and agent lifecycle — all under the Azure AI Foundry umbrella.

    In this video, we’ll break down:

    * What the Microsoft Agent Framework actually is — beyond the usual slides and headlines.

    * How MAF brings observability, governance, and responsible AI directly into the agent workflow.

    * The architectural stack powering MAF — from Azure AI Foundry to the developer toolchain.

    * And how it stands up against other frameworks like AutoGen, LangGraph, and Semantic Kernel — where it shines, and where it still has growing up to do.

    By the end, you’ll have a clear picture of why MAF is shaping up to be Microsoft’s playbook for building AI systems that aren’t just smart — they’re production-ready.

    Let’s get into it. ⚙️

    Thanks for reading! Subscribe for free to receive new posts and support my work.



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit manojknewsletter.substack.com
    Show More Show Less
    8 mins
  • Designing Limits that Scale — API Governance in Distributed Systems
    Nov 14 2025

    Welcome back to TechTalks with Manoj — the show where we go beyond buzzwords and break down the real architecture behind scalable, secure, and intelligent systems.

    Today, we’re talking about one of the most overlooked — yet absolutely critical — pillars of system design: API Rate Limiting and Traffic Management.

    It’s the invisible rulebook that keeps our systems fair, fast, and stable — even when the world hits “refresh” a million times a second.Most developers see rate limiting as a security feature. But for architects — it’s much more than that. It’s governance. It’s economics. It’s how we translate business contracts into system behavior.

    In this episode, we’ll explore:

    * How rate limiting evolved from a simple “safety brake” into a full-blown architectural control plane.

    * The algorithms that define fairness — from Token Buckets to Sliding Windows — and when to use each.

    * How distributed gateways coordinate global limits using Redis, Lua scripts, and consistent hashing.

    * Why infrastructure enforcement at the edge — through NGINX, Cloudflare, and API gateways — is the difference between resilience and chaos.

    * And how multi-tenant systems use rate limiting not just to protect themselves, but to enforce SLAs and even manage cost.

    By the end of this episode, you’ll understand that rate limiting isn’t about saying “no” — it’s about sustaining trust, performance, and fairness at scale.

    So if you’ve ever wondered why some APIs stay rock-solid under pressure while others crumble under traffic — this one’s for you.

    Let’s dive in. 🚦

    Thanks for reading! Subscribe for free to receive new posts and support my work.



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit manojknewsletter.substack.com
    Show More Show Less
    17 mins
No reviews yet
In the spirit of reconciliation, Audible acknowledges the Traditional Custodians of country throughout Australia and their connections to land, sea and community. We pay our respect to their elders past and present and extend that respect to all Aboriginal and Torres Strait Islander peoples today.