- The AI Pulse
- Posts
- š¤ OpenAI Releases āGPT-4.5ā
š¤ OpenAI Releases āGPT-4.5ā
PLUS: LLMs āThinkā Like Developers When Coding, New dLLMs Generate Over a 1,000 Tokens per Second

Welcome back AI enthusiasts!
In todayās Daily Report:
āļøOpenAI Releases āGPT-4.5ā
š§ LLMs āThinkā Like Developers When Coding
š¤ÆNew dLLMs Generate Over a 1,000 Tokens per Second
š Trending Tools
š„ŖBrief Bites
š°Funding Frontlines
š¼Whoās Hiring?
Read Time: 3 minutes
šRECENT NEWS
OPENAI
āļøOpenAI Releases āGPT-4.5ā

Image Source: Canvaās AI Image Generators/Magic Media
OpenAI released āGPT-4.5,ā the companyās largest and most knowledgeable AI model yet.
Key Details:
āGPT-4.5ā showcases better writing capabilities, improved world knowledge, and what OpenAI calls a ārefined personality.ā
OpenAI CEO Sam Altman explained that itās the first AI model that āfeels like talking to a thoughtful person to me.ā
OpenAI warned that āGPT-4.5ā isnāt a frontier AI model, but itās āOpenAIās largest LLM.ā
It was fine-tuned using Reinforcement Learning From Human Feedback (RLHF), which uses human feedback to teach LLMs to self-learn more efficiently and align with human preferences.
Itās currently available to ChatGPT Pro Plans and developers across all Paid API Tiers, with ChatGPT Plus, Team, and Enterprise Plans getting access next week.
Why Itās Important:
āGPT-4.5ā is super expensive to run, with developers across all Paid API Tiers paying 30x the input cost and 15x the output cost to use it.
āWeāre out of GPUs,ā said Altman. āWeāll add tens of thousands of GPUs next weekā¦This isnāt how we want to operate, but itās hard to perfectly predict growth surges that lead to GPU shortages.ā
š©ŗ PULSE CHECK
Are Tech Giants becoming overreliant on GPUs?Vote Below to View Live Results |
AI RESEARCH
š§ LLMs āThinkā Like Developers When Coding

Image Source: FAIR at Meta/āSWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolutionā/Screenshot
Metaās FAIR Team developed āSWE-RL,ā which enhances the reasoning abilities of LLMs to help them tackle real-world coding tasks.
Key Details:
āSWE-RLā relies on Reinforcement Learning (RL), which teaches LLMs to learn the optimal behavior in an environment to obtain the maximum reward.
RL is comprised of four components:
Learner: The LLMs
Environment: The real-world coding tasks the LLMs interact with.
Policy: The instructions the LLMs follow to take action.
Feedback: The Positive Rewards or Negative Penalties the LLMs observe after taking action.
Positive Rewards are given when the LLMs successfully write, debug, and test code.
Negative Penalties are given when the LLMs generate inefficient, error-prone code.
Why Itās Important:
When developers tackle real-world coding tasks, itās not just about writing code; itās about testing, debugging, and refactoring that code over and over.
āSWE-RLā helps LLMs not only generate code but also āthinkā like developers by constantly taking in feedback to adjust code.
INCEPTION LABS
š¤ÆNew dLLMs Generate Over a 1,000 Tokens per Second

Image Source: Inception Labs/āIntroducing Mercury, the first commercial-scale diffusion large language modelā/Screenshot
Inception Labs developed āMercury,ā a family of diffusion Large Language Models (dLLMs) that generate text faster than ever.
Traditional LLMs generate text from left to right, one Token at a time. In other words, a Token canāt be generated until all the text that comes before it has been generated.
Tokens are the smallest units of data used by LLMs to process and generate text. Similarly, we break down sentences into words or characters. You can think of Tokens as syllables. Simply put, Tokens represent bits of raw data; a million tokens equals roughly 750,000 words.
Instead of generating one Token at a time, dLLMs generate entire blocks of Tokens in parallel for increased speed, efficiency, and control. Itās 10x faster than traditional LLMs, 10x cheaper than traditional LLMs, and 2x the size of traditional LLMs with the same latency and cost.
š TRENDING TOOLS
šDeepTutor is your personalized AI tutor.
š¬Forage Mail declutters your email inbox.
š½ļøTopaz Labs brings old videos back to life.
š·Basalt integrates AI into your product in seconds.
š¬getcaramel.ai turns your ideas into revenue-generating ads.
š®Browse our always Up-To-Date AI Tools Database.
š„ŖBRIEF BITES
Hugging Face launched āFastRTC,ā an open-source Python library that helps developers build real-time audio and video AI apps.
Vevo Therapeutics created āTahoe-100M,ā the worldās largest single-cell dataset that maps out 60,000 drug-cell interactions.
NVIDIA CEO Jensen Huang said that nearly everyone would benefit from having a personalized AI-powered tutor with them at all times.
IBM unveiled āGranite 3.2,ā a family of small AI models that deploy Conditional Reasoning, Time Series Forecasting, and Document Vision to tackle Enterprise workloads.
š°FUNDING FRONTLINES
Hyperlume closes a $12.5M Seed Round to transform AI Data Center Connectivity.
Bridgetown Research secures a $19M Series A to build AI Agents for Enterprise Research.
Variational AI raises a $5.5M Seed Extension for AI-driven Small Molecule Drug Discovery.
š¼WHOāS HIRING?
Trepp (New York, NY): Data Science Intern, Summer 2025
CACI (Sterling, VA): Software Engineering Intern, Summer 2025
K2 Space (Los Angeles, CA): Loads and Dynamics Engineer, Entry-Level
Advarra (Remote): Data Scientist, Mid-Level
ThoughtSpot (Mountain View, CA): Senior Staff AI Architect, Senior-Level
šFINAL NOTE
FEEDBACK
How would you rate todayās email?It helps us improve the content for you! |
ā¤ļøTAIP Review of The Day
āsolid, holistic coverage of AI space! š«¶ā
REFER & EARN
šYour Friends Learn, You Earn!
You currently have 0 referrals, only 1 away from receiving āļøUltimate Prompt Engineering Guide.
Copy and paste this link to friends: https://theaipulse.beehiiv.com/subscribe?ref=PLACEHOLDER
Reply