Ask Runable forDesign-Driven General AI AgentTry Runable For Free
Runable
Back to Blog
Technology13 min read

xAI launches Grok 4.3 at an aggressively low price and a new, fast, powerful voice cloning suite | VentureBeat

The launch of Grok 4.3 represents a calculated bet by xAI that the market wants specialized brilliance and extreme cost efficiency over a perfectly balanced...

TechnologyInnovationBest PracticesGuideTutorial
xAI launches Grok 4.3 at an aggressively low price and a new, fast, powerful voice cloning suite | VentureBeat
Listen to Article
0:00
0:00
0:00

x AI launches Grok 4.3 at an aggressively low price and a new, fast, powerful voice cloning suite | Venture Beat

Overview

x AI launches Grok 4.3 at an aggressively low price and a new, fast, powerful voice cloning suite

While Elon Musk faces off against his former colleague and Open AI co-founder Sam Altman in court, Musk's rival firm x AI, founded to take on Open AI, isn't slowing down on launching competitive new products and services.

Details

Last night, x AI shipped a new, proprietary base large language model (LLM), Grok 4.3, and a new voice cloning suite on the web.

The new products arrive after months of tumult from x AI that saw all of Musk's 10 original co-founders of the lab and dozens more researchers exit the firm and Grok was eclipsed on performance by many new competing LLMs from the likes of Open AI, Anthropic, Google, and Chinese firms Deep Seek, Moonshot (Kimi), Alibaba (Qwen), z.ai, and others.

While Grok 4.3 does mark a significant leap in performance on third-party benchmarks over its direct predecessor Grok 4.2, according to the independent AI model evaluation firm Artificial Analysis, it still remains below the state-of-the-art set by Open AI and Anthropic's latest models.

But the marquee feature of the Grok brand has — other than Musk's stated opposition to "wokeness" and its more freewheeling personality and image generation policy — increasingly been its low price point when accessed by developers and users via the x AI application programming interface (API), a trend only furthered by Grok 4.3, which costs

1.25permillioninputtokensand1.25 per million input tokens and
2.50 per million output tokens (up to 200,000 input tokens, at which point costs double, a common pricing strategy of leading AI labs) compared to its direct predecessor Grok 4.2's initial API pricing of
2/2/
6 per million input/output tokens.

Grok 4.3 API pricing screenshot. Credit: Venture Beat

Grok 4.3 API pricing screenshot. Credit: Venture Beat

According to x AI's release notes, Grok 4.3 began beta testing in April for subscribers to x AI's Super Grok (

30monthly)plan,andthoseofitssiblingsocialnetwork,X,throughitsPremium+plan(30 monthly) plan, and those of its sibling social network, X, through its Premium+ plan (
40 monthly with 50% for first two months). Now it's available to all through the x AI API and through partner Open Router.

Reasoning baked-in and agentic tool-use capabilities

At the core of Grok 4.3 is a fundamental shift in how the model processes information. Unlike previous iterations where "chain-of-thought" or reasoning could often be toggled or configured by effort levels, Grok 4.3 is built with reasoning as an active, permanent state.

This means the model is designed to "think" before it speaks for every query, a strategy intended to maximize factual accuracy and the handling of complex, multi-step instructions.

The model’s memory is equally expansive, featuring a 1 million-token context window. To put this in perspective, a million tokens is roughly equivalent to several thick novels or the entire codebase of a mid-sized application.

This allows Grok 4.3 to maintain coherence over massive datasets, though x AI has implemented a "Higher context pricing" structure for requests that exceed the 200,000-token threshold.

This tiering suggests that while the "long-term memory" is available, the computational cost of managing that much information remains a significant overhead. Technically, the model accepts both text and image inputs, outputting text.

It is specifically optimized for agentic workflows—scenarios where an AI is not just answering a question but acting as an autonomous agent to complete a task.

For the first time, Grok has access to the same tools and environments a human professional would use. Evidence of this shift is visible in early user interactions:

Spreadsheet Engineering: In one instance, the model spent 6 minutes and 22 seconds in a "thought" phase to build a comprehensive OSRS Sailing Combat DPS analyzer. The resulting .xlsx file wasn't a simple table but a multi-sheet dashboard including a "Reference_Data" set and a complex "DPS_Calculator" with formulaic auto-calculations.

Spreadsheet Engineering: In one instance, the model spent 6 minutes and 22 seconds in a "thought" phase to build a comprehensive OSRS Sailing Combat DPS analyzer. The resulting .xlsx file wasn't a simple table but a multi-sheet dashboard including a "Reference_Data" set and a complex "DPS_Calculator" with formulaic auto-calculations.

Professional Documentation: Grok now generates formatted PDFs, such as 12-page reports on Space X products. These documents incorporate branding, logos, hero images, and structured tables, moving well beyond the markdown blocks of previous iterations.

Professional Documentation: Grok now generates formatted PDFs, such as 12-page reports on Space X products. These documents incorporate branding, logos, hero images, and structured tables, moving well beyond the markdown blocks of previous iterations.

Visual Presentations: The model can design 9-slide Power Point decks, utilizing a "Sandwich Structure" (dark titles/conclusions with light content) and integrating data-driven decision matrices and humor.

Visual Presentations: The model can design 9-slide Power Point decks, utilizing a "Sandwich Structure" (dark titles/conclusions with light content) and integrating data-driven decision matrices and humor.

However, its knowledge of the world is not infinite; the release notes list a knowledge cut-off date of December 2025. Yet, thanks to built-in web search, Grok can reference and use up-to-date information.

In fact, Grok 4.3 arrives with an enhanced ecosystem of tools designed to make it a functional digital employee. The x AI platform now offers a robust set of server-side tools that the model can invoke autonomously based on the complexity of the query.

Web and X Search: These tools allow Grok to bypass its knowledge cutoff by browsing the live internet or searching X (formerly Twitter) posts, user profiles, and threads.

Web and X Search: These tools allow Grok to bypass its knowledge cutoff by browsing the live internet or searching X (formerly Twitter) posts, user profiles, and threads.

Code Execution: The model can run Python code in a sandboxed environment to solve mathematical problems or process data.

Code Execution: The model can run Python code in a sandboxed environment to solve mathematical problems or process data.

File and Collections Search: A built-in Retrieval-Augmented Generation (RAG) system allows users to query uploaded document collections or search through specific file attachments.

File and Collections Search: A built-in Retrieval-Augmented Generation (RAG) system allows users to query uploaded document collections or search through specific file attachments.

x AI's Custom Voices let you clone your voice at high quality in a minute or two

Beyond text, x AI has introduced Custom Voices, a sophisticated voice-cloning API and web-based voice cloning creation suite.

This product allows developers to clone a voice from a reference audio clip as short as 120 seconds. Once cloned, the "voice ID" can be used across x AI’s Text-to-Speech (TTS) and Voice Agent APIs.

x AI's documentation emphasizes that this is not merely about timbre; the model is designed to pick up delivery patterns.

If a user records a reference clip in a "customer support" style, the resulting AI voice will mimic that helpful, professional inflection.

Despite the creative potential, x AI has placed strict geographic limits on this feature, making it available only in the United States, with a notable exception for Illinois due to regional biometric and privacy regulations.

While the console playground is open for general use, programmatic access via the POST /v 1/custom-voices endpoint is currently gated to teams on an Enterprise plan.

I tried it myself and after moving through the requisite voice sampling screens on the web — the tool asks you to read aloud several passages of unrelated dialog — I indeed had a copy of my voice that sounded eerily identical to mine and accurately pronounced new words the same way I would when reading allowed from a new script it was given.

You can delete your custom voices in one click on x AI's Custom Voices web application and create up to 30 new ones at a time.

Access to the new Voice Agent API (grok-voice-think-fast-1.0) is billed at a flat rate of

3.00perhour(3.00 per hour (
0.05 per minute) for speech-to-speech interactions. This is on the low-medium end of costs for other competing voice agents, according to my research:

Complementing this is the standalone Text-to-Speech (TTS) service, which offers five distinct voices (Eve, Ara, Rex, Sal, and Leo) and is priced at $4.20 per 1 million characters.

For transcription needs, the Speech-to-Text (STT) API provides real-time streaming at

0.20perhour,whilebatchprocessingisavailableatadiscountedrateof0.20 per hour, while batch processing is available at a discounted rate of
0.10 per hour.

To ensure security for client-side applications, x AI utilizes Ephemeral Tokens, allowing for secure Web Socket connections without exposing primary API keys.

Once created, these voices are private to the user's team and can be used across all voice APIs by referencing a unique 8-character alphanumeric voice_id.

For highly regulated sectors, x AI maintains production-ready standards, including SOC 2 Type II auditing, HIPAA eligibility for healthcare workloads, and GDPR compliance.

Aggressively low API pricing as a differentiator

The most aggressive aspect of the Grok 4.3 announcement is its pricing structure. Bindu Reddy, CEO of enterprise assistant startup Abacus AI noted on X that the model is "as smart as Sonnet 4.6 and 5x cheaper and faster".

The standard API rates are set at

1.25permillioninputtokensand1.25 per million input tokens and
2.50 per million output tokens. This reflects a significant reduction in cost compared to its predecessor, Grok 4.20, with Artificial Analysis reporting an approximately 40% lower input price and 60% lower output price.

According to our calculations at Venture Beat, that places Grok-4.3 firmly in the lowest cost half of all major foundation models, far closer to Chinese open source offerings than its U. S. proprietary rivals:

However, the "reasoning" nature of the model introduces a new billing category: Reasoning tokens.

These are tokens generated during the model's internal thinking process and are billed at the same rate as standard completion tokens. Effectively, users pay for the AI to "think" before it provides the final answer. x AI has also introduced several unique fee structures:

Prompt Caching: Repeated prompts are significantly cheaper, at $0.20 per million tokens, incentivizing developers to reuse context.

Prompt Caching: Repeated prompts are significantly cheaper, at $0.20 per million tokens, incentivizing developers to reuse context.

Tool Invocations: While token usage for tools is billed at standard rates, the act of invoking a tool carries a flat fee—

5.00per1,000callsforWebSearchorCodeExecution,and5.00 per 1,000 calls for Web Search or Code Execution, and
10.00 for File Attachments.

Tool Invocations: While token usage for tools is billed at standard rates, the act of invoking a tool carries a flat fee—

5.00per1,000callsforWebSearchorCodeExecution,and5.00 per 1,000 calls for Web Search or Code Execution, and
10.00 for File Attachments.

Usage Guideline Violation Fee: In a move that may set a new industry precedent, x AI charges a $0.05 fee for requests that are blocked by their safety filters before generation even begins.

Usage Guideline Violation Fee: In a move that may set a new industry precedent, x AI charges a $0.05 fee for requests that are blocked by their safety filters before generation even begins.

The model itself remains accessible via a standard commercial API, with x AI recommending that all developers migrate to grok-4.3 as their "most intelligent and fastest model".

The reception of Grok 4.3 has been polarized, depending largely on the specific use case. Professional benchmarkers and developers have highlighted a "stark gap" between the model's domain-specific strengths and its general reasoning consistency.

According to independent AI evaluation firm Vals AI, Grok 4.3 has taken the top spot on several specialized indices. It currently ranks #1 on Case Law v 2 (79.3% accuracy) and #1 on Corp Fin.

This 25-point jump in legal reasoning over Grok 4.20 suggests that the "always-on reasoning" architecture is particularly well-suited for the dense, logical structures of law and finance.

Artificial Analysis corroborated this performance, noting a massive improvement in agentic tasks, scoring an Elo of 1500 on the GDPval-AA benchmark, surpassing competitors like Gemini 3.1 Pro and GPT-5.4 mini.

Conversely, users focused on general-purpose agents and coding have highlighted deficiencies.

AI automated brick-and-mortar retail company Andon Labs reported that Grok 4.3 is a "big regression" on the Vending-Bench 2, which measures an AI's ability to take consistent actions in a simulation.

They colorfully described the model as having "narcolepsy problems," preferring to remain inactive for multiple simulation days rather than taking the required actions.

The sentiment was echoed by Vals AI, which noted that while the model improved in some coding areas, it remains weak on general coding tasks and "struggles with difficult math problems," scoring only 11% on Proof Bench.

The launch of Grok 4.3 represents a calculated bet by x AI that the market wants specialized brilliance and extreme cost efficiency over a perfectly balanced generalist.

By achieving a score of 53 on the Artificial Analysis Intelligence Index while remaining on the "Pareto frontier" of cost-per-intelligence, x AI is positioning itself as the "value" leader for enterprise applications in legal and financial tech.

The "always-on reasoning" is a double-edged sword. While it provides the depth needed to navigate complex case law, the community reports of "narcolepsy" suggest that a model that is always "thinking" may occasionally think itself into a state of paralysis, or at least a state of excessive caution that inhibits agentic action.

In addition, prior Grok model scandals including an X chatbot version referring to itself as "Mecha Hitler" and posting antisemitic content, sexualized deepfake imagery generation and investigations, and references to racial conflicts and right-wing dog whistle framing of social issues — which appear to mirror many of founder Musk's own positions, to the point that the model was at one point, checking Musk's own X account before responding in its X implementation — nearly certain to give some enterprises pause when considering adoption. It's unclear whether any of those issues remain with Grok 4.3, but one user did note that Grok's system prompt appears to instruct it "you do not assign broad positive/negative utility functions to groups of people."

For developers, the decision to adopt Grok 4.3 will likely come down to the nature of their data. For those needing to process a million tokens of legal documents at a fraction of the cost of Claude 4.6 or GPT-5.5, Grok 4.3 is a clear front-runner.

For those building high-frequency autonomous agents or complex math solvers, the "narcolepsy" and coding regressions suggest that x AI's latest model may still need a few more "tuning passes".

As Open Router noted on X upon making the model live, the "large jump in agentic performance" at a lower price point is an undeniable milestone. Whether that performance can be sustained across all domains remains the primary question for the summer of 2026.

Deep insights for enterprise AI, data, and security leaders

By submitting your email, you agree to our Terms and Privacy Notice.

Key Takeaways

  • x AI launches Grok 4

  • While Elon Musk faces off against his former colleague and Open AI co-founder Sam Altman in court, Musk's rival firm x AI, founded to take on Open AI, isn't slowing down on launching competitive new products and services

  • Last night, x AI shipped a new, proprietary base large language model (LLM), Grok 4

  • The new products arrive after months of tumult from x AI that saw all of Musk's 10 original co-founders of the lab and dozens more researchers exit the firm and Grok was eclipsed on performance by many new competing LLMs from the likes of Open AI, Anthropic, Google, and Chinese firms Deep Seek, Moonshot (Kimi), Alibaba (Qwen), z

Cut Costs with Runable

Cost savings are based on average monthly price per user for each app.

Which apps do you use?

Apps to replace

ChatGPTChatGPT
$20 / month
LovableLovable
$25 / month
Gamma AIGamma AI
$25 / month
HiggsFieldHiggsField
$49 / month
Leonardo AILeonardo AI
$12 / month
TOTAL$131 / month

Runable price = $9 / month

Saves $122 / month

Runable can save upto $1464 per year compared to the non-enterprise price of your apps.