MoreRSS

site iconTomasz TunguzModify

I’m a venture capitalist since 2008. I was a PM on the Ads team at Google and worked at Appian before.
Please copy the RSS to your reader, or quickly subscribe to:

Inoreader Feedly Follow Feedbin Local Reader

Rss preview of Blog of Tomasz Tunguz

Building Sales Teams in the Age of AI : Office Hours with Chris Klayko of Databricks

2025-10-07 08:00:00

Chris Klayko, SVP of Sales at Databricks

Chris Klayko brings over two decades of sales leadership experience transforming technology companies from promising startups to multi-billion dollar enterprises. As SVP of Sales at Databricks, Chris leads the charge in democratizing data & AI for organizations worldwide. His remarkable journey includes scaling Google Cloud from tens of millions to a multi-billion dollar business in just four years, driving UiPath’s Americas expansion during its hypergrowth phase, & building SAP’s emerging solutions division.

On October 15th in San Francisco, Theory Ventures is hosting an exclusive Office Hours session with Chris Klayko, SVP of Sales at Databricks.

During this intimate fireside chat, Chris & I will explore :

  • Building & motivating high-performing sales teams in the AI era
  • Lessons from scaling organizations from $100M → $1B+ at Databricks, UiPath & Google Cloud
  • How sales leaders can adapt in a fast-changing, AI-driven landscape
  • The evolution from traditional software sales to AI & data platform selling
  • Creating sales cultures that attract & retain top talent
  • Navigating the shift from product-led to enterprise sales motions

San Francisco | October 15, 5:30–7:30pm | Space is limited

Apply to join here. Submit your questions through the registration form & I’ll weave them into our conversation.

I look forward to welcoming Chris to Office Hours & sharing his insights with our community!

From Books to Prompts: The Instant Future of Productivity

2025-10-06 08:00:00

Last Saturday I spent three hours implementing a coding workflow I’d read about that morning. By Sunday afternoon I’d tested it on two projects , measured the results , & decided to keep it.

Ten years ago , the same journey would have taken weeks. Buy a Getting Things Done book , set up a Bullet Journal , or learn Zettelkasten. Read it. Understand the methodology. Build the habits. Iterate through mistakes. Maybe after a month you’d know if it worked.

Something fundamental shifted. We moved from commitment-based learning to experiment-based discovery. Three recent experiments show what’s possible.

First, Paper2Agent transforms research papers into working code agents in 30 minutes to 3 hours. Ask it to implement AlphaGenome’s genomic analysis from a published paper & it processes the code, builds the environment, runs the tutorials. Previously: read the paper (2 days), understand the methodology (1 day), code the implementation (3 days). Now: point AI at the repository, get a working agent by lunch.

Second, when I discovered Google’s Agent Development Kit research on tool design patterns, I asked AI to redesign my automation tools following those principles. The result : a 41% reduction in AI operation costs, implemented while I was answering emails in another browser tab.

Third, over the weekend I read Jesse Vincent’s architect/implementer workflow via Simon Willison’s write-up. The approach splits AI coding into two sessions: one architect to design, one implementer to execute. No weeks of trial & error to understand the nuances. Read, implement, test, done.

The pattern emerging : instant implementation.

We can now test productivity systems like trying coffee drinks. Monday : architect/implementer (cappuccino—structured , classic). Tuesday : traditional single-session (flat white—simpler , maybe better?). Wednesday : hybrid workflow (chai latte—wait , is this even coffee?). By Thursday , the data shows which saves 3 hours per week.

The self-help aisle promised transformation through commitment. AI promises transformation through experimentation.

Circular Financing: Does Nvidia's $110B Bet Echo the Telecom Bubble?

2025-10-03 08:00:00

When Nvidia announced a $100 billion investment commitment to OpenAI1 in September 2025 , analysts immediately drew comparisons to the telecom bubble. The concern : is this vendor financing , where a supplier lends money to customers so they can buy the supplier’s products , a harbinger of another spectacular collapse?

American tech companies will spend $300-400 billion on AI infrastructure in 20252,3 , exceeding any prior single-year corporate infrastructure investment in nominal dollars.3 David Cahn estimates the revenue gap has grown to $600 billion4.

I analyzed the numbers. The similarities are striking , but the differences matter.

The Lucent Playbook

Lucent vs Nvidia Revenue Comparison 1996-2024
Lucent’s revenue peaked at $37.92B in 1999 , crashed 69% to $11.80B by 2002 , never recovered. Merged with Alcatel in 2006.

In 1999 , Lucent Technologies reached $37.92 billion in revenue at the peak of the dot-com bubble. 5 Lucent was the #1 North American telecommunications equipment manufacturer with 157,000 employees & dominated markets alongside Nortel Networks (combined 53% optical transport market share). 6 Behind the scenes , equipment makers extended billions in vendor financing to telecom customers. Lucent committed $8.1B7 , Nortel extended $3.1B with $1.4B outstanding , & Cisco promised $2.4B in customer loans.8

The strategy seemed brilliant : lend money to cash-strapped telecom companies so they could buy your equipment. Everyone wins—until the merry-go-round stops.

When the bubble burst :

  • 47 Competitive Local Exchange Carriers (CLECs) bankrupted 2000-2003 , including Covad , Focal Communications , McLeod , Northpoint , Winstar 9,10
    • Why they failed : $60B overbuild 1996-2001 , market saturation from identical business models , sudden funding collapse (Jan 2001 : billions available , Apr 2001 : zero)11
  • 33-80% of vendor loan portfolios went uncollected as customers failed & equipment became worthless12
  • Fiber networks were using less than 0.002% of available capacity , with potential for 60,000x speed increases. 13 It was just too early.

Nvidia’s Playbook

Fast forward to 2025. Nvidia’s vendor financing strategy totals $110 billion in direct investments plus another $15+ billion in GPU-backed debt. The largest commitment is $100B to OpenAI (September 2025)1,14 , structured as 10 tranches of $10B each tied to infrastructure deployment milestones. The first $10B was valued at a $500B OpenAI valuation , with subsequent tranches priced at prevailing valuations. Payment comes via lease arrangements , not upfront GPU purchases. OpenAI CFO Sarah Friar confirmed : “Most of the money will go back to Nvidia”14

Beyond OpenAI , Nvidia holds a $3B stake in CoreWeave15 , a company that has spent $7.5B on Nvidia GPUs , & $3.7B in other AI startup investments16 through NVentures.

The GPU-backed debt market adds another layer. CoreWeave alone carries $10.45B in debt using GPUs as collateral17. An additional $10B+ in GPU-backed debt has emerged for “Neoclouds” including Lambda Labs ($500M GPU-backed loan)18,19.

Lucent in 1999-2000 had vendor financing commitments of $8.1B (24% of $33.6B revenue). Nvidia’s direct investments total 67% of annual revenue ($110B against $165B LTM). Nvidia’s exposure is 2.8x larger relative to revenue than Lucent’s official outstanding loans , though Lucent’s off-balance-sheet guarantees masked the true exposure.

The Numbers Side-by-Side (2024 Dollars)

Metric Lucent (FY2000, inflation-adj.) Nvidia (2025)
Vendor financing $15B $110B
Operating cash flow $304M20 $15.4B (Q2 FY26)
Revenue $61B $165B (LTM)
Top 2 Customers represent 23%21 39%

The Reasons to be Wary

1. The AI Customer Base is More Concentrated

Lucent’s top 2 customers—AT&T at 10% & Verizon at 13%—accounted for 23% of revenue in FY2000.21 The Regional Bell Operating Companies , or RBOCs , the seven “Baby Bells” created from AT&T’s 1984 breakup , were also major customers. Nvidia has 39% of revenue from just 2 customers & 46% from 4 customers , nearly double Lucent’s concentration. 88% of Nvidia’s revenue comes from data centers.

2. GPU-Backed Debt Is New

The new $10B+ GPU-backed debt market is built on the assumption that GPUs will hold their value over 4-6 years. GPU-backed loans carry ~14% interest rates22 , triple investment-grade corporate debt.23

How Depreciation Schedules Changed :

Company Pre-2020 2020-2021 2022-2023 2024-2025 Change
Amazon24 3 years 4 years (2020) → 5 years (2021) 5 years 6 years (2024) → 5 years (2025) First reversal
Microsoft25 ~3 years 4 years 6 years 6 years +100%
Google26 ~3 years 4 years 6 years 6 years +100%
Meta27 ~3 years 4 years 4.5 years → 5 years 5.5 years +83%
CoreWeave28 N/A N/A 4 years → 6 years (Jan 2023) 6 years +50% (GPUs)
Nebius29 N/A N/A 4 years 4 years Industry standard

Amazon’s 2025 reversal (6 → 5 years) is the first major pullback.

CPUs historically have 5-10 years of useful life , while GPUs in AI datacenters last 1-3 years in practice , despite 6-year accounting assumptions.30,31 Evidence from Google architects shows GPUs at 60-70% utilization survive 1-2 years , with 3 years maximum.31 Meta’s Llama 3 training experienced 9% annual GPU failure rates , suggesting 27% failure over 3 years.31

Cerno Capital raises the question : “Are these policies a reflection of genuine economic & technological realities? Or are these policies a lever by which hyperscalers are enhancing the optics of their investment programs amid rising investor concerns?”32

4. The Use of SPVs

Tech companies use Special Purpose Vehicles (SPVs) to finance AI datacenter construction. A hyperscaler like Meta partners with a private equity firm like Apollo , contributing capital to a separate legal entity that builds & owns the datacenter.

As investor Paul Kedrosky explains : “I have a stake in it as Meta. Some giant private debt provider has a stake in it. The datacenter is under my control. But I don’t own it, so you don’t get to roll it back into my balance sheet.”2*

The Structure

  1. Entity Creation : Hyperscaler & PE firm form separate legal entity (SPV)
  2. Capital Structure : Typically 10-30% equity, 70-90% debt from private credit markets
  3. Lease Agreement : SPV leases capacity back to hyperscaler
  4. Balance Sheet Treatment : SPV debt doesn’t appear on hyperscaler’s balance sheet

The hyperscaler maintains operational control through long-term lease agreements. Because it doesn’t directly own the SPV , the debt remains off its balance sheet under current accounting standards.

The appeal is straightforward. “I don’t want the credit rating agencies to look at what I’m spending. I don’t want investors to roll it up into my income statement.”2*

Market Scale

American tech companies are projected to spend $300-400 billion on AI infrastructure in 2025. Hyperscaler capital expenditures have reached approximately 50% of operating income2, levels historically associated with government infrastructure buildouts rather than technology companies.

Where the Risk Sits

Datacenter assets now represent 10-22% of major REIT portfolios2 , up from near zero two years ago. The thin equity layer (10-30%) means if datacenter utilization falls short of projections or if GPUs depreciate faster than projected , equity holders face losses before debt holders experience impairment.

*Quotes lightly edited for clarity & brevity

5. Custom Silicon Threat

Hyperscalers are building their own AI accelerators to reduce Nvidia dependence. Microsoft aims to use “mainly Microsoft silicon” , specifically Maia accelerators , in datacenters.33 Google deploys TPUs , Amazon builds Trainium & Inferentia chips , & Meta develops MTIA processors. If customers shift to in-house silicon , CoreWeave’s GPU collateral value & Nvidia’s vendor financing become exposure to customers building competitive alternatives.

Nvidia Isn’t Lucent & 2025 Isn’t 2000

  • Accounting : Lucent manipulated $1.148B in revenue , SEC charged 10 executives with fraud5 ; Nvidia shows no evidence of manipulation , audited by PwC , Aa3 rated34
  • Cash flow : Lucent lent $8.1B while cash flow lagged profitability & receivables exploded $5.4B (1998-1999)20 ; Nvidia lends with $50B+ annual operating cash flow & $46.2B net cash35
  • Credit rating : Lucent downgraded to A3 (December 2000)36 ; Nvidia upgraded to Aa3 (March 2024)34
  • Customer base : Lucent’s customers were leveraged CLECs burning capital ; Nvidia’s top 4 customers generated $451B in operating cash flow in 2024 (Microsoft $119B , Alphabet $125B , Amazon $116B , Meta $91.3B)37
  • Capacity : Fiber networks used <0.002% of capacity in 200013 ; Microsoft & AWS report AI capacity constraints in 202538,39

What I’m Watching

Is AI demand real (like cloud computing) or speculative (like dot-com fiber)?

Here’s what I’m watching :

  1. GPU utilization rates : Are data centers actually using the chips or just stockpiling?
  2. OpenAI’s monetization : Can they generate enough revenue to justify the buildout?
  3. Debt defaults : Any cracks in the $15B GPU-backed debt market?
  4. AR trends : AR improved from 68% (FY24) to 30% (Q2 FY26) , but still watch for deterioration
  5. Customer adds : Are new customers emerging , or is Nvidia dependent on the same 2-4 hyperscalers?
  6. Custom silicon threat : Microsoft developing Maia accelerators , aiming to use “mainly Microsoft silicon in the data center.”33 If hyperscalers shift to in-house chips , Nvidia’s vendor financing becomes exposure to customers building competitive alternatives.
  7. Vendor consolidation : Many companies are in a period of experimentation trying 2-3 competing vendors. Those experimental budgets may thin with time , reducing overall spend.

AI is already broadly deployed—40% of US employees used AI at work by September 2025 , double the 20% rate in 2023.40 Questions persist about effectiveness : the oft cited MIT study found 95% of AI pilots failed to deliver measurable P&L impact , primarily due to poor integration rather than technical failures.41

Yet the pace of improvement is tremendous. Labor market data shows wages rising twice as fast in AI-exposed industries , & workers using AI boost performance up to 40%.40 Many of Nvidia’s customers are profitable & sophisticated hyperscalers—Microsoft , Google , Amazon , Meta—generating $451B in operating cash flow in 2024 , with tremendous pull from their own enterprise customers demanding AI. OpenAI is not profitable , reporting a $4.7B loss in H1 2025 on $4.3B revenue , though nearly half the loss is stock-based compensation.42

Unlike the telecom bubble , where demand was speculative & customers burned cash , this merry-go-round has paying riders.


Coda : Lucent’s Accounting Fraud

Behind the vendor financing disaster was systematic accounting fraud. The SEC charged Lucent with manipulating $1.148 billion in revenue & $470 million in pre-tax income during fiscal year 2000. 5 The fraud involved multiple schemes :

Channel Stuffing : Lucent sent $452 million in equipment to distributors but counted it as revenue before the distributors sold to end customers.5 This created phantom sales.

Side Agreements : Lucent executives entered secret agreements with distributors granting them return rights & privileges beyond their distribution contracts , making it improper to recognize revenue.5 These side deals were hidden from auditors.

Reserve Manipulation : Lucent improperly established & maintained excess reserves to smooth earnings , violating GAAP.5

The SEC charged 10 Lucent executives with securities fraud.5 The company paid a $25 million fine—the largest ever for failing to cooperate with an SEC investigation.5 The accounting manipulation masked deteriorating fundamentals until too late.

The WinStar Collapse : Lucent committed $2 billion in vendor financing to WinStar Communications , a CLEC. When WinStar struggled , Lucent refused a final $90 million loan extension. WinStar filed bankruptcy. Lucent wrote off $700 million in bad debts.43 This pattern repeated across customer defaults : Lucent made provisions for bad debts of $2.2 billion (2001) & $1.3 billion (2002)—a total of $3.5 billion in customer loan losses.43


References


  1. “Nvidia to Invest Up to $100 Billion in OpenAI”, CNBC (September 22, 2024) ↩︎ ↩︎

  2. Paul Kedrosky , “This Is How the AI Bubble Could Burst”, Plain English with Derek Thompson podcast (September 23, 2024) ; “SPVs, Credit & AI Datacenters”, Paul Kedrosky blog ↩︎ ↩︎ ↩︎ ↩︎ ↩︎

  3. “OpenAI, Oracle, and SoftBank expand Stargate”, Stargate $500B commitment details ↩︎ ↩︎

  4. “AI’s $600B Question”, Sequoia Capital analysis by David Cahn showing AI revenue gap expanded from $125B to $600B ↩︎

  5. Lucent Technologies financial data & accounting fraud details ↩︎ ↩︎ ↩︎ ↩︎ ↩︎ ↩︎ ↩︎ ↩︎

  6. “Nortel Networks and Lucent Technologies dominate North American optical transport market”, Lightwave (1999) ; “Who Lost Lucent?”, American Affairs Journal confirming Lucent $41.4B revenue fiscal 2000 & combined 53% market share ↩︎

  7. Lucent vendor financing commitments ↩︎

  8. “Cisco, Lucent & Nortel: Prime Lenders for Network Buildout”, TheStreet (2001) ↩︎

  9. Industry analysis of telecom bankruptcies 2000-2003 , including 47 CLEC failures ↩︎

  10. “Competitive Local Exchange Carrier”, Wikipedia ↩︎

  11. “The rise and fall of the competitive local exchange carriers in the U.S.”, ResearchGate academic analysis ; “The Great Telecom Implosion”, Princeton analysis ↩︎

  12. Industry analysis of vendor financing losses during telecom bubble collapse ↩︎

  13. Fiber Broadband Association , “Fiber Broadband Scalability & Longevity” white paper (2024) ; IEEE research on optical fiber capacity ↩︎ ↩︎

  14. “Nvidia’s investment in OpenAI will be in cash, and most will be used to lease Nvidia chips”, CNBC interview with OpenAI CFO Sarah Friar (September 24, 2024) ↩︎ ↩︎

  15. Industry reports on CoreWeave equity & GPU purchases (2024) ↩︎

  16. Nvidia investor presentations & NVentures portfolio data (2024) ↩︎

  17. “CoreWeave Raises $7.5 Billion in Debt”, Bloomberg (May 2024) ↩︎

  18. Lambda Labs GPU financing announcements (2024) ↩︎

  19. Financial Times reporting on GPU-backed debt market emergence (2024) ↩︎

  20. Lucent cash flow vs net income analysis ↩︎ ↩︎

  21. Lucent Technologies 10-K Annual Report (FY2000) : “Revenues from AT&T accounted for approximately 10% of consolidated revenues in fiscal 2000. Revenues from Verizon accounted for approximately 13% of consolidated revenues in fiscal 2000.” ↩︎ ↩︎

  22. “CoreWeave’s GPU-Backed Debt Strategy”, AIinvest analysis of ~14% interest rates on GPU-backed loans ↩︎

  23. “As venture debt gambles on GPUs, not all are sold on silicon-backed loans”, PitchBook analysis of GPU collateral risks ↩︎

  24. “Amazon Revises Server Lifespan” & The Register reporting on AWS depreciation changes (2020-2025) ↩︎

  25. “Accounting for AI: Hyperscaler Depreciation Policies”, Cerno Capital analysis ↩︎

  26. “Google Extends Server Life to Six Years”, Data Center Dynamics (2023) ↩︎

  27. “Meta Extends Server Life”, The Stack Technology (2025) ↩︎

  28. “CoreWeave Depreciates Its GPUs Over 6 Years”, WCCFtech (2025) ↩︎

  29. “How Long Do GPUs Last Anyway?”, Applied Conjectures analysis of GPU depreciation policies ↩︎

  30. “Datacenter GPU service life can be surprisingly short — only one to three years”, Tom’s Hardware ; “How Long Should a GPU Actually Last?”, confirming 3-4 year GPU lifecycle vs CPUs at 5-7 years ↩︎

  31. “Datacenter GPU service life can be surprisingly short”, Tom’s Hardware reporting on Google architect analysis ↩︎ ↩︎ ↩︎

  32. “Accounting for AI: Financial Accounting Issues and Capital Deployment in the Hyperscaler Landscape”, Cerno Capital analysis (2025) ↩︎

  33. “Microsoft wants to use ‘mainly Microsoft silicon’ in its data centers”, The Register (October 2, 2025) ↩︎ ↩︎

  34. Moody’s Investors Service upgrade of Nvidia to Aa3 (March 2024) ↩︎ ↩︎

  35. Nvidia Q2 FY26 Financial Results (ended July 27, 2025) ↩︎

  36. Lucent credit rating downgrades ↩︎

  37. Fiscal year 2024 operating cash flow data from company financial statements : Microsoft FY2024 (ended June 30, 2024) Form 10-K , Alphabet FY2024 (ended December 31, 2024) Form 10-K , Amazon FY2024 (ended December 31, 2024) Form 10-K , Meta FY2024 (ended December 31, 2024) $91.328B operating cash flow ↩︎

  38. "$13b Run Rate & Doubling", Tomasz Tunguz analysis of Microsoft Q3 2025 earnings (January 30, 2025) ↩︎

  39. “Google’s Future in Search & AI”, Tomasz Tunguz analysis citing AWS capacity constraints (2025) ↩︎

  40. “Anthropic Economic Index report: Uneven geographic and enterprise AI adoption”, Anthropic (September 2025) showing 40% of US employees used AI at work , double the 20% in 2023 ; “AI in Productivity: Top Insights and Statistics for 2024”, showing workers using AI boost performance up to 40% & wages rising twice as fast in AI-exposed industries ↩︎ ↩︎

  41. “MIT report: 95% of generative AI pilots at companies are failing”, Fortune (August 2025) ; Study by Aditya Challapally found 95% of AI pilots failed to deliver measurable P&L impact , primarily due to poor integration with existing workflows rather than technical AI model failures ↩︎

  42. “OpenAI’s First Half Results: $4.3 Billion in Sales, $2.5 Billion Cash Burn”, The Information ; OpenAI reported $4.3B revenue & $4.7B loss in H1 2025 , with stock-based compensation expenses approaching $2.5B , nearly half the total loss ↩︎

  43. WinStar bankruptcy & Lucent bad debt provisions ↩︎ ↩︎

Data &amp; AI Infrastructure Are Fusing

2025-10-02 08:00:00

Screenshot 2025-10-02 at 11.11.44 AM

AI breaks the data stack.

Most enterprises spent the past decade building sophisticated data stacks. ETL pipelines move data into warehouses. Transformation layers clean data for analytics. BI tools surface insights to users.

This architecture worked for traditional analytics.

But AI demands something different. It needs continuous feedback loops. It requires real-time embeddings & context retrieval.

Consider a customer at an ATM withdrawing pocket money. The AI agent on their mobile app needs to know about that $40 transaction within seconds. Data accuracy & speed aren’t optional.

Netflix rebuilt their entire recommendation infrastructure to support real-time model updates1. Stripe created unified pipelines where payment data flows into fraud models within milliseconds2.

The modern AI stack requires a fundamentally different architecture. Data flows from diverse systems into vector databases, where embeddings & high-dimensional data live alongside traditional structured data. Context databases store the institutional knowledge that informs AI decisions.

AI systems consume this data, then enter experimentation loops. GEPA & DSPy enable evolutionary optimization across multiple quality dimensions. Evaluations measure performance. Reinforcement learning trains agents to navigate complex enterprise environments.

Underpinning everything is an observability layer. The entire system needs accurate data & fast. That’s why data observability will also fuse with AI observability to provide data engineers & AI engineers end-to-end understanding of the health of their pipelines.

Data & AI infrastructure aren’t converging. They’ve already fused.


References


  1. Netflix Technology Blog. (2025, August). “From Facts & Metrics to Media Machine Learning: Evolving the Data Engineering Function at Netflix.” https://netflixtechblog.com/from-facts-metrics-to-media-machine-learning-evolving-the-data-engineering-function-at-netflix-6dcc91058d8d ↩︎

  2. Stripe. (2025). “How We Built It: Stripe Radar.” https://stripe.com/blog/how-we-built-it-stripe-radar ↩︎

Adding Complexity Reduced My AI Cost by 41%

2025-09-30 08:00:00

I discovered I was designing my AI tools backwards.

Here’s an example. This was my newsletter processing chain : reading emails, calling a newsletter processor, extracting companies, & then adding them to the CRM. This involved four different steps, costing $3.69 for every thousand newsletters processed.

Before: Newsletter Processing Chain

# Step 1: Find newsletters (separate tool)
ruby read_email.rb --from "[email protected]" --limit 5
# Output: 340 tokens of detailed email data

# Step 2: Process each newsletter (separate tool)
ruby enhanced_newsletter_processor.rb
# Output: 420 tokens per newsletter summary

# Step 3: Extract companies (separate tool)
ruby enhanced_company_extractor.rb --input newsletter_summary.txt
# Output: 280 tokens of company data

# Step 4: Add to CRM (separate tool)
ruby validate_and_add_company.rb startup.com
# Output: 190 tokens of validation results

# Total: 1,230 tokens, 4 separate tool calls, no safety checks
# Cost: $3.69 per 1,000 newsletter processing workflows

Then I created a unified newsletter tool which combined everything using the Google Agent Development Kit, Google’s framework for building production grade AI agent tools :

# Single consolidated operation
ruby unified_newsletter_tool.rb --action process \
  --source "techcrunch" --format concise \
  --auto-extract-companies
# Output: 85 tokens with all operations completed

# 93% token reduction, built-in safety, cached results
# Cost: $0.26 per 1,000 newsletter processing workflows
# Savings: $3.43 per 1,000 workflows (93% cost reduction)

Why is the unified newsletter tool more complicated?

It includes multiple actions in a single interface (process, search, extract, validate), implements state management that tracks usage patterns & caches results, has rate limiting built in, & produces structured JSON outputs with metadata instead of plain text.

But here’s the counterintuitive part : despite being more complex internally, the unified tool is simpler for the LLM to use because it provides consistent, structured outputs that are easier to parse, even though those outputs are longer.

To understand the impact, we ran tests of 30 iterations per test scenario. The results show the impact of the new architecture :

Metric Before After Improvement
LLM Tokens per Op 112.4 66.1 41.2% reduction
Cost per 1K Ops $1.642 $0.957 41.7% savings
Success Rate 87% 94% 8% improvement
Tools per Workflow 3-5 1 70% reduction
Cache Hit Rate 0% 30% Performance boost
Error Recovery Manual Automatic Better UX

We were able to reduce tokens by 41% (p=0.01, statistically significant), which translated linearly into cost savings. The success rate improved by 8% (p=0.03), & we were able to hit the cache 30% of the time, which is another cost savings.

While individual tools produced shorter, “cleaner” responses, they forced the LLM to work harder parsing inconsistent formats. Structured, comprehensive outputs from unified tools enabled more efficient LLM processing, despite being longer.

My workflow relied on dozens of specialized Ruby tools for email, research, & task management. Each tool had its own interface, error handling, & output format. By rolling them up into meta tools, the ultimate performance is better, & there’s tremendous cost savings. You can find the complete architecture on GitHub.

From A to B Without Inventing Letters

2025-09-29 08:00:00

“The way to do a piece of writing is three or four times over, never once.”

Writing is hard. John McPhee, who invented literary nonfiction that reads like a novel, developed a four-draft writing method that transforms chaotic ideas into compelling narratives.

McPhee pioneered creative nonfiction at The New Yorker, writing books like Oranges & Coming into the Country that made complex subjects fascinating through storytelling. His approach differs from traditional journalism by incorporating fiction techniques while maintaining factual accuracy. His prose combines vivid imagery with economy :

“The doctor listens in with a stethoscope and hears sounds of a warpath Indian drum.”

He favored directness :

“He liked to go from A to B without inventing letters between.”

About his genre, McPhee said :

“Nonfiction—what the hell, that just says, this is nongrapefruit we’re having this morning.”

McPhee later codified his approach in Draft No. 4: On the Writing Process, sharing decades of writing wisdom.

His organizational philosophy shapes everything :

“You can build a structure in such a way that it causes people to want to keep turning pages. A compelling structure in nonfiction can have an attracting effect analogous to a story line in fiction. Readers are not supposed to notice the structure. It is meant to be about as visible as someone’s bones.”

McPhee’s Four-Draft Framework :

  1. Brain dump draft - Capture every possible idea, fact, & angle without editing or judgment
  2. Structure draft - Organize ideas into logical sequences & identify the core narrative thread
  3. Ruthless cut draft - Remove everything that doesn’t serve the primary message or confuse the reader
  4. Polish draft - Refine prose, fix grammar, & ensure each sentence drives toward your goal

This is one of the best techniques I’ve found for writing. The method works because it separates creative thinking from critical evaluation. When you try to write perfect prose while generating ideas, it’s easy to fall into creative block.

Each draft becomes the foundation for the next, creating a recursive process that transforms chaotic thoughts into structured narratives. Like peeling back the layers of an orange to reveal the fruit within, each draft strips away what doesn’t belong, revealing the essential story that was always there waiting to be discovered.