๐ŸŒFreshcollected in 63m

Fractile raises $220m for in-memory-compute inference chip production

Fractile raises $220m for in-memory-compute inference chip production
PostLinkedIn
๐ŸŒRead original on The Next Web (TNW)

๐Ÿ’กNew hardware architecture aiming to solve the memory bottleneck for LLM inference with major industry backing.

โšก 30-Second TL;DR

What Changed

Secured $220 million in funding led by Accel.

Why It Matters

This funding signals a shift toward specialized hardware architectures that bypass the memory wall, potentially offering significant latency improvements for large-scale LLM inference.

What To Do Next

Monitor Fractile's public benchmarks against H100s to evaluate if their in-memory architecture fits your specific inference workload requirements.

Who should care:Developers & AI Engineers

๐Ÿง  Deep Insight

Web-grounded analysis with 12 cited sources.

๐Ÿ”‘ Enhanced Key Takeaways

  • โ€ขFractile was founded in 2022 by Walter Goodwin, an Oxford PhD in robotics, who developed the concept while researching large language models (LLMs) for general-purpose robots.
  • โ€ขThe company projects its chips to deliver AI inference performance that is 100 times faster, 10 times cheaper, and 20 times more energy-efficient than current Nvidia GPUs, specifically for LLMs like Llama2-70B.
  • โ€ขFractile's in-memory compute architecture utilizes SRAM (Static Random-Access Memory) to integrate memory and compute directly on the same die, thereby mitigating the data transfer bottleneck prevalent in conventional GPU-DRAM systems.
  • โ€ขThe startup emerged from stealth in July 2024 with an initial $15 million seed funding round and has committed ยฃ100 million to expand its UK operations, including establishing a new hardware engineering facility in Bristol.
  • โ€ขThe recent $220 million funding round, co-led by Factorial Funds, Accel, and Peter Thiel's Founders Fund, reportedly values Fractile at over $1 billion.

๐Ÿ› ๏ธ Technical Deep Dive

<ul><li>Fractile's core technology is an in-memory compute architecture designed for AI inference, particularly for large language models (LLMs).</li><li>The architecture integrates compute and memory directly on the same silicon die.</li><li>It employs SRAM (Static Random-Access Memory) to co-locate memory and processing units, aiming to eliminate the 'memory wall' or 'data-shuttling bottleneck' associated with moving data between traditional GPUs and off-chip DRAM.</li><li>This approach is projected to achieve a 100-fold increase in effective bandwidth and significantly higher energy efficiency.</li><li>Fractile claims its accelerators can run LLMs like Llama2-70B 100 times faster, at one-tenth the system cost, and 20 times more energy-efficiently than Nvidia H100 GPUs for decode tokens per second.</li><li>The company is developing custom multiply-accumulate (MAC) circuits that also store state.</li><li>There is speculation that Fractile may optimize MAC arrays for general matrix-vector (GEMV) operations rather than general matrix-matrix (GEMM) operations to enhance efficiency.</li><li>The Fractile team includes experienced engineers from companies such as Graphcore, Nvidia, and Imagination Technologies.</li><li>Fractile is developing its own software stack in conjunction with its hardware.</li><li>Commercial readiness for its chips is anticipated around 2027.</li></ul>

๐Ÿ”ฎ Future ImplicationsAI analysis grounded in cited sources

Fractile's technology could significantly reduce the operational costs and energy consumption of large-scale AI inference.
By integrating compute and memory on a single die using SRAM, Fractile aims to eliminate the memory bottleneck, leading to substantial improvements in speed, cost, and energy efficiency compared to traditional GPU architectures.
The emergence of companies like Fractile will intensify competition in the AI inference chip market, potentially diversifying the supply chain beyond dominant players like Nvidia.
Fractile's innovative approach and significant funding, coupled with Anthropic's reported interest in diversifying its chip suppliers, indicate a growing market for specialized inference hardware that challenges existing solutions.
Fractile's success could accelerate the development and deployment of more complex 'reasoning models' in AI.
Pat Gelsinger noted that reasoning models are memory-bound and require generating thousands of output tokens, a limitation Fractile's in-memory compute aims to overcome, enabling faster and more efficient execution of such advanced AI.

โณ Timeline

2022
Fractile founded by Walter Goodwin.
2024-07
Fractile emerged from stealth and announced $15 million in seed funding.
2024-10
Received a $6.52 million grant from the UK government's ARIA program.
2025-01
Pat Gelsinger announced his investment in Fractile.
2026-02
Fractile announced plans to invest ยฃ100 million to expand UK operations, including a new hardware engineering facility in Bristol.
2026-05
Secured $220 million in funding led by Accel, Factorial Funds, and Founders Fund, valuing the company at over $1 billion.

๐Ÿ“Ž Sources (12)

Factual claims are grounded in the sources below. Forward-looking analysis is AI-generated interpretation.

  1. vertexaisearch.cloud.google.com โ€” Auziyqgjaf1sg9ziaqtijejkl7xf E Qm0hofn N7htoyzgqilkywclh8zljpettpggljsyl Rdbglrskcvkmzpl6f Efhs511fgcyethpqjb2lt5obm7 Lpmn4nceqmkrndawlmz 0j1fzlbdqiqlqcu66vxlkmawoheuadx9rbkz6zhvtmofqozkzfanorirl3jeyrvdmsxvphc A3epsbfewj09ybmpbs1xtgbuykivg8wskzrw==
  2. vertexaisearch.cloud.google.com โ€” Auziyqg0cdbmuvmyn0q4mzrkdd3k Gr L Nvycskqvev4oiuktkyfrqfvjgpad1k38uvwwd5zcs7k6p8ehtu4eydsdqxjtsf V8z7x1wtmpxssz0uwxvwut8krfooeti9u7hruqmtd2t66gc4nfzvx8mykkm8agja00okfddfyudkxlfpz 0ukywiz8blmqiuwe7mmkgq49g0wlryuxmb8xepjf9
  3. vertexaisearch.cloud.google.com โ€” Auziyqe66srd24e9o Arswbj0ejibqw2oq Mu Kty4g7ftzohiyrzaieye8b9ptszgsgjcmwz U3hucvr80q8txk to Sem3quqthj2lpyerxrafwto17w8iasapinzimdxuid8ulusqysr14qr7ndzwehfpxc9v68jc7z9ittj1sj2rf U095ellc7bcno3ewr5xemuyte Ckzw6c9vxh2ldq2zdorwpcfuxfaobp Wvmbinsspve2oruxao Z R0tjsxkxfea2x6js
  4. vertexaisearch.cloud.google.com โ€” Auziyqg6faz Fzi Jrrivomx3xmw4thtfkwkw2eznd5ab9ehdvl8x2nnt3mvtpemcigoact5q668reoi4eoa7qsjnwhsg Haidzx3imjeeikkywmgjjg2makxlmy6rs6209dighcfubif4fdvtskkn6ukkw Xqieziej8ogumaj Ktoogysdxjtvl Wy0a==
  5. vertexaisearch.cloud.google.com โ€” Auziyqefigogay Ya4nu3qdgmszdf41l6ndenuddurd0nrxxlxtp Zm8oata6jxpaipmmgxjnpsps Qn9ys4rih4fkmjxpf24ul6mugjdg 8wmx Inhh4pitb Vmbyzheat1fycm6w=
  6. vertexaisearch.cloud.google.com โ€” Auziyqgkjzjmvkd2wmq7yma6mukxnzu T Vczmdvf2h8r3s 76wybvma3gmiuxdf7tp9eklb7wr0nzwxffcau9repzelxtdgugv3thqjwp Zmxvcljfcbcro8aekedkvtyrnvdisd1buwbz7fvu1vtnm0xyzts8wc7tjno6kjxdwnngntzpyloija==
  7. vertexaisearch.cloud.google.com โ€” Auziyqe4onau45lbqhi7hjqg Zuvdnpwyjepub 825rgbg7qyqyvyfkxxakhadtmbjis9qj58flmr3zfok5cdsuknn0oikigthp1idow94njdh6huq6akbqt64m7nmwepc46bukx8pnaxhgm3cnkmdcq5tksl Fvmi2jdnstqjdvdkzvzoajoemj0hbofo Pulf5liwgyl3egb Oaulonqjkmy1 Kw==
  8. vertexaisearch.cloud.google.com โ€” Auziyqhn6g1oqysaiwrg7kusgywpfcnacz5jxvdaaxi3oyaxgk8ghqtcgngpxu59s7hfi8hjchsu7pxdgnzuchndmvnu7cmti4ofwreopwhs93htodruuzjk Tb1wouvvfbgorpckw==
  9. vertexaisearch.cloud.google.com โ€” Auziyqhqw7obdcg2gmas42ninahe6s2xdwrf5pxswqzsdyhwovbwndfqytwk8ifdj V Urt9jpi2c9cyeivf2fo6025y3dxp Ypfdzucpge3aaqradfalxsom5bm5q66e Nbepuvb9omvjyvi9uxaerzo2hp1 Bduybcfm6lzwowhxia6w8uwo0umqndbjq=
  10. vertexaisearch.cloud.google.com โ€” Auziyqh6maey67xyiu0 Qa6vbfdofrh6wv 6dz0iwgnn9duoq6xyr3mj4ze7vzjxuwj9amvbkaibxpbcxl3vlub5clk Zjrpbc4agcw Odgecuxldoulsuc08a2jdat3oshlsiuiyufcca0aia0pgm3v9nlwlnerrbyn4htkdeq Stxkdomvca==
  11. vertexaisearch.cloud.google.com โ€” Auziyqgpnplff Ejagwwwrs23r0j9fmbkmk5dp2i8uect4o0t4sxmfv9jq8m9lqhqepo5zs9upouy Tzwrrlyvkb9nfkdradkmi P0lw9lllqskrxe5v E8y2zat1j2ayn3lg K=
  12. vertexaisearch.cloud.google.com โ€” Auziyqeqg9dlqi 1d3viq Udpkhlds2nnjz4puyhgmztub5pkhm9o34acb1lxsgcqeryld6vbdlnzwxajplz7j2yq6pb By1o0w8cutrlxjo8l5r Cc Gnivbj71polm Dx2zn7ilsohjmvhfc5t Shi Mrsy 1033lnwzamuc2uhdqbfkx4qk7txbt5jpiozwtnkltvk1ufcmbwnbv9sraajxpner4=
๐Ÿ“ฐ

Weekly AI Recap

Read this week's curated digest of top AI events โ†’

๐Ÿ‘‰Related Updates

AI-curated news aggregator. All content rights belong to original publishers.
Original source: The Next Web (TNW) โ†—