Topics

belated

AI

Amazon

Article image

Image Credits:Jaap Arriens/NurPhoto(opens in a new window)/ Getty Images

Apps

Biotech & Health

Climate

Cloud Computing

DoC

Crypto

endeavor

EVs

Fintech

fundraise

gizmo

Gaming

Google

Government & Policy

Hardware

Instagram

Layoffs

Media & Entertainment

Meta

Microsoft

Privacy

Robotics

security measures

Social

Space

Startups

TikTok

transfer

Venture

More from TechCrunch

Events

Startup Battlefield

StrictlyVC

Podcasts

video

Partner Content

TechCrunch Brand Studio

Crunchboard

Contact Us

Elon Musk ’s AI startup , X.ai , has let out its late generative AI theoretical account , Grok-1.5 . Set to power social connection X ’s Grok chatbot in the not - so - aloof time to come ( “ in the come days , ” per ablog post ) , Grok-1.5 is likely a mensurable upgrade over its forerunner , Grok-1 — at least pronounce by the published bench mark results and eyeglasses .

Grok-1.5 benefits from “ improved reasoning , ” according to X.ai , peculiarly where it concerns dupe and math - bear on labor . The model more than double Grok-1 ’s score on a democratic math benchmark , MATH , and scored over 10 pct points higher on the HumanEval test of programming language coevals and problem - figure out abilities .

It ’s difficult to predict how those results will translate in factual usage . As we recentlywrote , commonly - used AI benchmark , which valuate things as esoteric as performance on alum - level chemical science exam motion , do a poor caper of get how the mediocre soul interacts with manikin today .

One improvement thatshouldlead to evident amplification is the amount of context Grok-1.5 can realise compared to Grok-1 .

Grok-1.5 can sue contexts of up to 128,000 tokens . Here , “ tokens ” name to morsel of in the raw text ( for instance , the parole “ fantastical ” split into “ fan , ” “ tas ” and “ tic ” ) . Context , or linguistic context window , refers to stimulus data ( in this case , textual matter ) that a model reckon before generate output ( more text ) . Models with small setting window tend to bury the contents of even very late conversations , while models with prominent contexts avoid this pit — and , as an added welfare , full savvy the flow of data point they take in .

“ [ Grok-1.5 can ] utilize selective information from considerably longsighted documents , ” X.ai write in the web log post . “ Furthermore , the model can handle longer and more complex command prompt while still uphold its statement - following capacity as its context windowpane expands . ”

What ’s historically set X.ai ’s Grok posture apart from other generative AI model is that they answer to doubt about topics that aretypically off - boundary to other model , like confederacy and more controversial political ideas . The fashion model also answer questions with “ a rebellious streak , ” as Musk has described it , and unlimited rude oral communication if bespeak to do so .

Join us at TechCrunch Sessions: AI

Exhibit at TechCrunch Sessions: AI

It ’s ill-defined what changes , if any , Grok-1.5 impart in these areas . X.ai does n’t advert to this in the blog billet .

Grok-1.5 will soon be usable to early testers on X , go with by “ several new features . ” Musk has antecedently hint at summarizing threads and response , and suggesting contentedness for military post ; we ’ll see if those arrive soon enough .

The announcement comes after X.aiopen sourcedGrok-1 , albeit without the code necessary to fine - tune or further check it . More recently , Musk said that more users on disco biscuit — specifically those bear for X ’s $ 8 - per - calendar month Premium plan — would earn access to the Grok chatbot , which was previously only available to disco biscuit Premium+ client ( who pay $ 16 per month ) .