<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Aaron's Insights]]></title><description><![CDATA[AI Startup Advisor at Amazon. Early-Stage Product & GTM. Free Forever. Opinions are my own. ]]></description><link>https://blog.aaronamelgar.me</link><image><url>https://substackcdn.com/image/fetch/$s_!g-1X!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45443bcb-1dd2-479f-b378-ae6c98655e77_758x758.png</url><title>Aaron&apos;s Insights</title><link>https://blog.aaronamelgar.me</link></image><generator>Substack</generator><lastBuildDate>Sun, 03 May 2026 01:34:28 GMT</lastBuildDate><atom:link href="https://blog.aaronamelgar.me/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Aaron Melgar]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[aaronamelgar@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[aaronamelgar@substack.com]]></itunes:email><itunes:name><![CDATA[Aaron Melgar]]></itunes:name></itunes:owner><itunes:author><![CDATA[Aaron Melgar]]></itunes:author><googleplay:owner><![CDATA[aaronamelgar@substack.com]]></googleplay:owner><googleplay:email><![CDATA[aaronamelgar@substack.com]]></googleplay:email><googleplay:author><![CDATA[Aaron Melgar]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Test Time Scaling with Human Feedback]]></title><description><![CDATA[Integrating Human Feedback into AI Reasoning Pipelines and implications on UX]]></description><link>https://blog.aaronamelgar.me/p/test-time-scaling-with-human-feedback</link><guid isPermaLink="false">https://blog.aaronamelgar.me/p/test-time-scaling-with-human-feedback</guid><dc:creator><![CDATA[Aaron Melgar]]></dc:creator><pubDate>Wed, 05 Mar 2025 20:48:03 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e041bdac-d28f-4fce-b5ae-4bfe9bf8516b_800x800.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I doubt I'm the only person wondering about this - but when will the test time scaling paradigm incorporate human feedback throughout the reasoning pipeline?</p><p>For those of you who aren't familiar with test time scaling, the latest reasoning systems including DeepSeek R1, Open AI o3, and Stanford's recent <a href="https://arxiv.org/pdf/2501.19393">s1</a> are all built on this architecture. In this framework, prompts don't just generate a single response immediately, but instead invoke a series of models to respond to the prompt and each output is then evaluated and reasoned on by yet another model to choose the best output before proceeding to the next step. These intermediate reasoning steps are performed in parallel or sequentially, sometimes a combination of both. At each reasoning step a variety of methods, such as majority voting, decides which outputs to propagate to the next step. These reasoning steps are typically evaluated by models sometimes known as "Models as a Judge", and more advance techniques include building reinforcement learning reward functions to provide feedback to the models in the previous step. Checkout this <a href="https://venturebeat.com/ai/how-test-time-scaling-unlocks-hidden-reasoning-abilities-in-small-language-models-and-allows-them-to-outperform-llms/">article</a> for a simplified breakdown if you want to go deeper.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.aaronamelgar.me/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Aaron's Insights! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In the math domain, these reasoning steps are easily verifiable and proceed down logical paths. This is why so many of the benchmarks for reasoning are all evaluating how well the model does on very difficult math problems like <a href="https://epoch.ai/frontiermath">FrontierMath</a>. In others like coding, reasoning steps are verifiable as well, e.g. did the code compile - however I'd argue coding is subjective, there are many ways to solve a problem dependent on the programmers style, context, or intent. For generalized questions and research, which are frankly more practical for most applications than PhD-level math, reasoning may not be verifiable at all and even more subject to the user preference.</p><p>I'm not ML researcher, but I'd suspect that its possible for these systems to solicit human feedback on its intermediate reasoning steps in real time. Of course the evaluator model needs be able to decide when to ask for human feedback - but assuming that's possible, this should provide better results. In my experience, the o family of models do a decent job of asking for reasoning feedback up front, but imagine a world where the model pauses and checks its work with you. A lot of people say AI is like a junior employee, and when you work with junior employees you give them feedback on work in progress not just their output.</p><p>I want to also explore the UX implications of test time scaling with human feedback. The psychology of the human computer interface has trended faster and faster since the origins of "instant message" chat. In some cases like real-time voice, companies like Cartesia are pushing the boundaries of physics where latency truly matters for a good UX. In many other use cases especially with text, I don't think low latency and complete autonomy is as important as people say it is, we're just biased to expect that. In my experience, users are actually eager to provide feedback and steer AI systems rather than getting frustrated by low quality outputs. Looking at products like Cursor, part of its success stems from people being able to specify the context the model should use, rather than just hoping the backend vector search system can figure out which files to use.</p><p>As test time scaling advances, these workloads are going to become increasingly asynchronous going from seconds to hours or even days for a single query. If we shift the AI UX paradigm to how delegate asynchronous workflows today, we can easily picture an AI system Slacking or emailing you asking for feedback on its reasoning. Taking this further, an AI workflow could even set the status of a Linear ticket to 'needs review' when people should share code feedback before going back to work. Designing the UX for asynchronous workloads also helps bring down the cost of AI by allowing cheaper or time-shared computational resources to run these reasoning pipelines. This greatly extends budgets, allowing models to take more steps to reason on a problem.</p><p>Looking forward to seeing how UX for AI evolves as we delegate more complex tasks to test time scaling systems. If anyone is working on this I'd love to chat!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.aaronamelgar.me/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Aaron's Insights! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[IQh - a new way of pricing AI systems]]></title><description><![CDATA[Explore into a new unit of measure for pricing AI reasoning inspired by physics.]]></description><link>https://blog.aaronamelgar.me/p/iqh-a-new-way-of-pricing-ai-systems</link><guid isPermaLink="false">https://blog.aaronamelgar.me/p/iqh-a-new-way-of-pricing-ai-systems</guid><dc:creator><![CDATA[Aaron Melgar]]></dc:creator><pubDate>Thu, 30 Jan 2025 00:17:39 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ab6b80b5-f679-4458-9a4b-6d6883970b42_800x800.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As a physics major, I distinctly recall my professors stressing the importance of getting the units of measure correct. Even if the final number was off by a factor of pi, oftentimes having the appropriate units and being within the right order of magnitude earned me at least partial credit on exams. Looking ahead into the future of AI has me pondering the question: are tokens the right units to price AI models?</p><p>Take a simple example:</p><p>&#8220;Is 7428958698238956023987723967931 prime?&#8221; - roughly 20 tokens</p><p>&#8220;Take this hour long call transcript, plus this 30 page compliance guide and tell me if the sales person asked about pre-existing conditions&#8220; - roughly 30,000 tokens.&nbsp;</p><p>Should these input and output tokens be priced the same? Clearly a model system will apply more effort answering one prompt over the other. </p><p>In these simple examples, high-level model tiers like 2.5, Pro, or Lite may suffice. However this starts to break down as model systems become increasingly dynamic and involve many inference steps to generate answers. While most are touting 2025 as the year of Agents, one of the most important trends is the rise of chain-of-thought reasoning systems, which involve multiple parallel model inferences to generate, critique, refine, and reason about a prompt. DeepSeek&#8217;s <a href="https://github.com/deepseek-ai/DeepSeek-R1">R1 reasoning mode</a>l took the world by storm, open sourcing the architecture and model  - making reasoning capabilities significantly cheaper. They weren&#8217;t the first to build an architecture in this direction though, OpenAI&#8217;s<a href="https://www.forrester.com/blogs/openais-o3-hype-or-a-real-step-toward-agi/"> o-3</a> unveiling just at the end of last year sparked a huge debate around the future AI system design striving towards general intelligence. Meta also just released a paper outlining a reasoning system called<a href="https://arxiv.org/html/2412.06769v1"> COCONUT</a> which uses embeddings rather than text at intermediate steps to improve performance and efficiency. Even early-stage startups like Fireworks have been building model systems for <a href="https://fireworks.ai/blog/fireworks-compound-ai-system-f1">complex reasoning</a>, and no doubt this system design pattern will continue to evolve well into 2025 and beyond.</p><p>One could argue that chain-of-thought reasoning is itself an Agentic pattern, but I&#8217;m not here to play semantics. Autonomously applying reinforcement learning and supervised-fine-tuning on systems of models at scale seems beyond the scope of database queries and customer service workflows that most people refer to as Agents but I digress. Pricing these reasoning systems requires a slew of new economic  considerations that encapsulate how much time and effort went into generating an answer.</p><ul><li><p>Number of refinement steps generating the answer</p></li><li><p>Number of refinement steps occurring in parallel</p></li><li><p>Size of the models at each step</p></li><li><p>Latency requirements</p></li><li><p>Quantity of reasoning tokens exchanged between steps</p></li><li><p>Amount of context pulled from retrieval or search systems</p></li><li><p>So many more&#8230;&nbsp;</p></li></ul><p>As the levers of control with these systems become increasingly granular, I would bet that engineering teams will be able to provision reasoning capabilities with a certain degree of accuracy and precision. &#8220;I want a system with an IQ of 100 to work on this problem for up to 4 hours&#8221; more accurately represents how people think about labor economics. When you think about it, this is actually how people price consulting services - a higher hourly rate for a more capable person.&nbsp;</p><p>These reasoning systems can even dynamically route a prompt to an appropriately powerful model, and even generate reasoning plans within budget constraints. The reasoning capabilities of a model combined with the amount of effort to arrive at answer seems to be the best way to price these systems. </p><p>This led me to the notion of IQ-hour or IQh which succinctly captures the interplay between reasoning capability and effort applied. Of course the measure of IQ doesn&#8217;t come without its limitations and caveats. The relationship between IQ and socioeconomic status is highly contested, and IQ doesn&#8217;t capture domain specific performance. In my head, IQ is distinct from domain knowledge which comes in the form of context, like pulling medical research into a prompt. The model system then reasons on the body of knowledge with varying degrees of IQ. All of this said, IQ is the most widely accepted unit for quantifying intelligence and reasoning, and makes for a good shorthand approximation here.&nbsp;</p><p>An analogous unit to IQh in physics is the kWh unit of energy which is how your electricity usage is billed, where kW represents the power generated multiplied by time consumed in hours. The simple example would be with 10wH, you can turn on a 1W light for 10 hours, or turn on a more powerful 10W light for 1h. In this case, IQ represents the reasoning power of the model, multiplied by the time it gets used. A powerful model would solve a task in less time but could still cost the same as a less powerful model taking longer. time. After all, Andrew Ng famously proclaimed that &#8220;<a href="https://www.gsb.stanford.edu/insights/andrew-ng-why-ai-new-electricity">AI is the new electricity</a>&#8221; back in 2017, great to see this coming to fruition.&nbsp;</p><p>I could see an argument that IQh is simply an extension of a GPU-hour metric, and that each step in the chain of thought reasoning can be priced by the amount of time the computer is being utilized. However I tend to favor output pricing over input pricing; your utility doesn&#8217;t bill for the natural gas consumed or cost of solar installations, they just charge you for the kWh you use. If you have two competing reasoning offerings, each could price at the same GPU-hour level but have different outcomes based on each system design. True outcome based pricing in these systems is intractable because of their general nature, and determining the value of outcomes is often an arduous task. Just ask your an engineer about the value of addressing technical debt and you&#8217;re in for hours of debate..&nbsp;</p><p>IQ is also more fungible as a potential standardized definition of reasoning, whereas it goes without saying that not all high performance compute clusters are created equal. As Nvidia releases new chipsets every year, AMD and others race to compete, and many cloud providers are<a href="https://semianalysis.com/2024/12/03/amazons-ai-self-sufficiency-trainium2-architecture-networking/"> provisioning their own chips</a>. It&#8217;s easier to consider lowering the cost of IQh, than increasing the reasoning capability of a GPU-hr because compute clusters are so heterogeneous and sensitive to the economics of purchasing the hardware. Each cloud provider procures and provisions with their own methodology, but the output unit of IQ can be applied ubiquitously across any application.</p><p>I think IQh concept is well aligned with the economics and heuristics of building AI systems. Even if we don&#8217;t get down to IQ precision of measuring reasoning, some combination of reasoning power and time is the right direction. In the spirit of strong opinions, weakly held - I&#8217;d love to hear how other people are thinking about pricing AI reasoning systems.</p>]]></content:encoded></item></channel></rss>