{"id":8049,"date":"2025-08-26T10:19:00","date_gmt":"2025-08-26T10:19:00","guid":{"rendered":"https:\/\/sandbox.hbmadvisory.com\/amplify\/the-hidden-costs-of-ai-publishing-why-you-need-to-pick-the-right-model\/"},"modified":"2025-08-26T10:47:08","modified_gmt":"2025-08-26T10:47:08","slug":"the-hidden-costs-of-ai-publishing-why-you-need-to-pick-the-right-model","status":"publish","type":"post","link":"https:\/\/sandbox.hbmadvisory.com\/amplify\/the-hidden-costs-of-ai-publishing-why-you-need-to-pick-the-right-model\/","title":{"rendered":"The hidden costs of AI publishing: why you need to pick the right model"},"content":{"rendered":"<p><\/p>\n<div>\n<p>When ChatGPT-5 launched, I was ready to double down on my loyalty to OpenAI. Tokens looked cheaper on paper, and, as a publisher-facing tech business, NoahWire consumes AI at scale to create, curate and distribute content. Every improvement in the models means an improvement in what we can deliver. But within weeks that optimism turned into hard questions about cost, speed and reliability.<\/p>\n<p>On launch day, Sam Altman, OpenAI\u2019s ceo, proudly said API usage had \u201cdoubled overnight\u201d. That puzzled me. Publishers, for example, don\u2019t suddenly double the number of calls they make to an API just because a model gets smarter \u2013 their daily requirements are tied to editorial workflows.\u00a0<\/p>\n<p>What I soon realised was that while usage hadn\u2019t changed, costs had. Token burn shot up. Even Altman, who I admire and often think of as the closest thing we have to a Messiah in this industry, seemed to have missed this. He believed people were simply using the system more, when in fact the model itself was consuming more tokens to do the same work.<\/p>\n<p><strong>Paying more for the same job<\/strong><\/p>\n<p>GPT-5 tokens may be cheaper per unit, but the model is verbose and \u201cover thinks\u201d even simple editorial tasks. A summarisation or headline job that GPT-4 completed neatly now consumes far more tokens. Worse, it runs slower.\u00a0<\/p>\n<p>For publishers who integrate AI into live pipelines, where feeds, newsletters, CMS integrations and alerts demand split-second accuracy, unpredictability is as damaging as under-performance. Even \u201cover-performance\u201d can cause problems when you\u2019re building precisely tuned stacks.<\/p>\n<p>Then there are the online costs. Models like Gemini tempt publishers with a free allowance, but once you\u2019re running tens of thousands of queries a day \u2013\u00a0the norm for live feeds \u2013\u00a0the bills explode. OpenAI\u2019s trick is subtler: GPT-5 looks cheaper per unit, but by bloating responses and tokens it drives up the true cost in production. Hidden charges like these can quietly wreck a publishing P&amp;L.<\/p>\n<p><strong>Rethinking model choices<\/strong><\/p>\n<p>This has pushed us, and the publishers we work with, to rethink how tasks are assigned across models:<\/p>\n<p>* ChatGPT-5 \u2013 Still the best for polished, multilingual long-form drafts. Excellent for features or in-depth explainers, but too slow and costly for daily pipelines\n<\/p>\n<p>* GPT-4 \u2013 Cheaper, faster and more predictable. Less \u201cbrilliant\u201d than GPT-5, but a reliable workhorse for summaries, headlines and bulletins\n<\/p>\n<ul>\n<li>Gemini \u2013 Strong technically, but the cost of online calls makes it impractical for publishers running live news or alerts\n<\/li>\n<li>DeepSeek \u2013 Initially impressive, but quickly became unreliable in tests, drifting mid-task. There\u2019s also unease about how it sourced its training data. I invest in China myself, but in AI publishing it can feel like swimming in shark-infested waters.<br \/>\n<\/li>\n<\/ul>\n<p><strong>What this means for publishers<\/strong><\/p>\n<p>ChatGPT-5 was supposed to simplify publishing operations. Instead, it has pushed publishers \u2013\u00a0and many others beyond publishing \u2013\u00a0into becoming experts in model economics, workflow design and cost optimisation. Entire weeks have been lost re-engineering systems simply to stabilise them.<\/p>\n<p>Perhaps this is the new reality of AI publishing. The winners won\u2019t be those who pledge loyalty to one model, but those who treat AI as a toolkit, choosing the right model for each editorial job and constantly auditing cost against output.<\/p>\n<p>For publishers, that means high-quality features may still be written with ChatGPT-5, but live newsrooms, newsletters and bulletins demand the speed and economy of simpler models. The true cost of AI publishing isn\u2019t just about tokens \u2013\u00a0it\u2019s about online calls, latency and efficiency. Those who grasp this will publish faster, cheaper and better. Those who don\u2019t will keep getting caught out.<\/p>\n<p><em>Ivan Massow is founder and ceo of <a href=\"https:\/\/noahwire.com\/\" rel=\"nofollow noopener\" target=\"_blank\">NoahWire<\/a><\/em><\/p>\n<\/p><\/div>\n","protected":false},"excerpt":{"rendered":"<p>When ChatGPT-5 launched, I was ready to double down on my loyalty to OpenAI. Tokens looked cheaper on paper, and, as a publisher-facing tech business, NoahWire consumes AI at scale to create, curate and distribute content. Every improvement in the models means an improvement in what we can deliver. But within weeks that optimism turned<\/p>\n","protected":false},"author":1,"featured_media":8050,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[118],"tags":[],"class_list":{"0":"post-8049","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-publishing-news"},"amp_enabled":true,"_links":{"self":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts\/8049","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/comments?post=8049"}],"version-history":[{"count":1,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts\/8049\/revisions"}],"predecessor-version":[{"id":8051,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts\/8049\/revisions\/8051"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/media\/8050"}],"wp:attachment":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/media?parent=8049"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/categories?post=8049"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/tags?post=8049"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}