In-depth analysis of decentralized AI reasoning: How to strike a balance between security and performance | Bee Network
True “decentralization” Blockchain and Proof of Reasoning How it actually works Security and performance trade-offs Optimize space When we started building Gonka , we had a vision: What if anyone could run AI inference and get paid for it? What if we could harness all that unused computing power instead of relying on expensive centralized providers? The current AI landscape is dominated by a handful of large cloud providers: AWS, Azure, and Google Cloud control the majority of the world’s AI infrastructure. This centralization creates serious problems that many of us have experienced firsthand. Control of AI infrastructure by a handful of companies means they can arbitrarily set prices, censor undesirable applications, and create a single point of failure. When OpenAI’s API went down, thousands of applications crashed with it. When AWS experienced an outage, half the internet stopped functioning. Even “efficient” cutting-edge technologies aren’t cheap. Anthropic previously stated that training Claude 3.5 Sonnet cost “tens of millions of dollars,” and while Claude Sonnet 4 is now generally available, Anthropic hasn’t yet released its training costs. Its CEO, Dario Amodei, previously predicted that training costs for cutting-edge models will approach $1 billion, with the next wave of models reaching billions. Running inference on these models is equally expensive. For a moderately active application, a single LLM inference run can cost hundreds to thousands of dollars per day. Meanwhile, the world has a vast amount of computing power sitting idle (or being used in meaningless ways). Think of Bitcoin miners burning electricity to solve worthless hash puzzles, or data centers running below capacity. What if this computing power could be used for something truly valuable, like AI inference? A decentralized approach can pool computing power, lowering capital barriers and reducing single-supplier bottlenecks. Instead of relying on a few large companies, we can create a network where anyone with a GPU can participate and get paid for running AI inference. We know that building a viable decentralized solution will be incredibly complex. From consensus mechanisms to training protocols to resource allocation, there are countless pieces that need to be coordinated. Today, I want to focus on just one aspect: running inference on a specific LLM . Just how difficult is that? What is true decentralization? When we talk about decentralized AI inference, we mean something very specific. It’s not just about having AI models running on multiple servers, but about building a system where anyone can join, contribute computing power, and be rewarded for honest work. The key requirement is that the system must be trustless . This means you don’t have to trust any single person or company to run the system correctly. If you’re letting strangers on the internet run your AI model, you need cryptographic guarantees that they’re actually doing what they claim to be doing (at least with a sufficiently high probability). This trustless requirement has some interesting implications. First, it means the system needs to be verifiable : you need to be able to prove that the same model and the same parameters were used to generate a given output. This is particularly important for smart contracts that need to verify that the AI responses they receive are legitimate. But there’s a challenge: the more verification you add, the slower the entire system becomes, as network power is consumed by verification. If you completely trust everyone, there’s no need for verification reasoning, and performance is almost identical to centralized providers. But if you trust no one and always verify everything, the system becomes incredibly slow and uncompetitive with centralized solutions. This is the core contradiction we have been working to resolve: finding the right balance between security and performance . Blockchain and Proof of Reasoning So how do you actually verify that someone ran the correct model and parameters? Blockchain becomes an obvious choice — while it has its own challenges, it remains the most reliable way we know of to create an immutable record of events. The basic idea is fairly straightforward. When someone runs inference, they need to provide proof that they used the correct model. This proof is recorded on the blockchain, creating a permanent, tamper-proof record that anyone can verify. The problem is that blockchains are slow. Really, really slow. If we tried to record every step of reasoning on-chain, the sheer volume of data would quickly overwhelm the network. This constraint drove many of our decisions when designing the Gonka Network. When designing a network and thinking about distributed computing, there are multiple strategies to choose from. Can you shard a model across multiple nodes, or keep the entire model resident on a single node? The primary limitations come from network bandwidth and blockchain speed. To make our solution feasible, we chose to fit a full model on a single node, though this may change in the future. This does impose a minimum requirement for joining the network, as each node needs sufficient computing power and memory to run the entire model. However, a model can still be sharded across multiple GPUs belonging to the same node, giving us some flexibility within the constraints of a single node. We use vLLM, which allows customization of tensor and pipeline parallelism parameters for optimal performance. How it actually works Therefore, we agreed that each node hosts a complete model and runs full inference, eliminating the need for coordination across multiple machines during the actual computation. The blockchain is used only for record-keeping. We only record transactions and artifacts used for inference verification. The actual computation occurs off-chain. We want the system to be decentralized, without any single central point directing inference requests to network nodes. In practice, each participant deploys at least two nodes: a network node and one or more reasoning (ML) nodes. The network nodes are responsible for communication (including a chain node connecting to the blockchain and an API node managing user requests), while your ML nodes perform LLM inference. When an inference request arrives on the network, it reaches one of the API nodes (acting as a “transfer agent”), which randomly selects an “executor” (an ML node from a different participant). To save time and parallelize blockchain logging with the actual LLM computation, the transfer agent (TA) first sends the input request to the executor and records the input on-chain while the executor’s ML node is running inference. Once the computation is complete, the executor sends the output to the TA’s API node, while its own chain node records a verification artifact on-chain. The TA’s API node transmits the output back to the client, which is also recorded on-chain. Of course, these records still contribute to overall network bandwidth constraints. As you can see, blockchain recording neither slows down the start of the inference computation nor the time it takes for the final result to be returned to the client. Verification of whether the inference was completed honestly occurs later, in parallel with other inferences. If the executor is caught cheating, they lose the entire epoch’s reward, and the client is notified and receives a refund. The final question is: What is included in the artifact, and how often do we verify our reasoning? Security and performance trade-offs The fundamental challenge is that security and performance are at odds with each other. If you want maximum security, you need to verify everything. But that’s slow and expensive. If you want maximum performance, you need to trust everyone. But that’s risky and opens you up to all sorts of attacks. After some trial, error, and parameter tuning, we found an approach that attempted to balance these two considerations. We had to carefully tune the amount of verification, the timing of verification, and how to make the verification process as efficient as possible. Too much verification, and the system becomes unusable; too little verification, and the system becomes insecure. Keeping the system lightweight is crucial. We maintain this by storing the top k next- token probabilities. We use these to measure the likelihood that a given output was indeed generated by the claimed model and parameters, and to capture any tampering attempts, such as using a smaller or quantized model, with sufficient confidence. We will describe the implementation of the inference verification procedure in more detail in another post. At the same time, how do we decide which inferences to verify and which not? We chose a reputation-based approach. When a new participant joins the network, their reputation is 0, and 100% of their inferences must be verified by at least one participant . If a problem is found, the consensus mechanism will ultimately determine whether your inference is approved, or your reputation will be lowered, and you may be kicked off the network. As your reputation grows, the number of inferences that need verification decreases, and eventually 1% of inferences may be randomly selected for verification. This dynamic approach allows us to keep the overall verification percentage low while effectively catching participants who attempt to cheat. At the end of each epoch, participants are rewarded in proportion to their weight in the network. Tasks are also weighted, so rewards are expected to be proportional to both weight and the amount of work completed. This means we don’t need to catch and punish cheaters immediately; it’s sufficient to catch them within the epoch before distributing rewards. Economic incentives drive this trade-off as much as technical parameters. By making cheating expensive and honest participation profitable, we can create a system where the rational choice is honest participation. Optimize space After months of building and testing, we’ve built a system that combines the record-keeping and security advantages of blockchain while approaching the single-shot inference performance of centralized providers. The fundamental tension between security and performance is real, and there’s no perfect solution, only different trade-offs. We believe that as the network scales, it has a real opportunity to compete with centralized providers while maintaining full decentralized community control. There’s also significant room for optimization as it develops. If you’re interested in learning about this process, please visit our GitHub and documentation, join the discussion on Discord, and participate in the network yourself. About Gonka.ai
Gonka is a decentralized network designed to provide efficient AI computing power. Its design goal is to maximize the use of global GPU computing power to complete meaningful AI workloads. By eliminating centralized gatekeepers, Gonka provides developers and researchers with permissionless access to computing resources while rewarding all participants with its native GNK token. Gonka was incubated by US AI developer Product Science Inc. Founded by the Libermans siblings, Web 2 industry veterans and former core product directors at Snap Inc., the company successfully raised $18 million in 2023 from investors including OpenAI investor Coatue Management, Solana investor Slow Ventures, K5, Insight, and Benchmark Partners. Early contributors to the project include well-known leaders in the Web 2-Web 3 space, such as 6 Blocks, Hard Yaka, Gcore, and Bitfury. Official Website | Github | X | Discord | Whitepaper | Economic Model | User Manual This article is sourced from the internet: In-depth analysis of decentralized AI reasoning: How to strike a balance between security and performanceRecommended Articles Related: Prediction Market New Infrastructure: An Inventory of Polymarket’s Third-Party Ecosystem Projects Polymarket has become almost synonymous with prediction markets. Public information shows that Polymarket’s monthly trading volume has repeatedly surpassed $1 billion this year, creating a significant gap between it and second-place Kalshi. The project has not only received tens of millions of dollars in investment from Trump’s eldest son, but is also preparing to return to the US market and secure a new round of funding. Market speculation suggests its valuation could reach $10 billion. Against this backdrop, a number of third-party ecosystems have emerged around Polymarket, encompassing data/dashboards, social experiences, front-end/terminals, insurance, and AI agents. On September 12th, RootData compiled a collection of representative projects into its “Polymarket Ecosystem Projects” compilation, which we will explore in this article. Polysights | One-stop analysis panel Polysights is a one-stop analytics dashboard… # Analysis# bitcoin# crypto# Token© Copyright NoticeThe copyright of the article belongs to the author, please do not reprint without permission. Pre The Rise of a New RWB Narrative: From “Asset Tokenization” to “Business Tokenization” Next Plasma's Stablecoin Dream Story: The 26-Year-Old Founder Tells How to Build the Strongest Project of This Round Related articles HotViewpoint: Crypto narrative has entered the post-performance era, and over-optimizing infrastructure performance is mean 6086cf14eb90bc67ca4fc62b 39,807 10 With the wealth effect seriously lost, can Ethereum survive the midlife crisis? 6086cf14eb90bc67ca4fc62b 36,191 1 There are AI Agent projects everywhere. How to judge whether Dev is reliable through GitHub? 6086cf14eb90bc67ca4fc62b 37,566 2 Robinhood opened a big game late at night, kicking out the imagination of exchanges throughout the summer 6086cf14eb90bc67ca4fc62b 28,643 3 Hyperliquid, a new player in the cryptocurrency stock market 6086cf14eb90bc67ca4fc62b 29,959 Is the end of traditional banks approaching? Tether’s $500 billion valuation challenges OpenAI. 6086cf14eb90bc67ca4fc62b 18,744 2 No comments You must be logged in to leave a comment! Login immediately No comments... Latest Articles How to Systematically Track High-Win-Rate Addresses on Polymarket? 5hrs ago 255 CoinEx Research: Geopolitical Tensions Drive Up Oil and Gold Prices, Crypto Market Absorbs Liquidity Shock 5hrs ago 393 Low-Threshold Investment in SpaceX and ByteDance: MSX Partners with Republic to Usher in a New Era of Global Top Unicorn Investment 5hrs ago 363 Hold Bitcoin Mid-Term Short Positions, HYPE Successfully Rides the Wave for Profits | Guest Analysis 5hrs ago 402 Arthur Hayes: Middle East Flares Up, Time to Be Bullish on Bitcoin 5hrs ago 432 Popular WebsitesTempoLighterGAIBGliderPlanckRaylsBCPokerVooi Bee.com The world's largest Web3 portal Partners CoinCarp Binance CoinMarketCap CoinGecko Coinlive Armors Download Bee Network APP and start the web3 journey White Paper Roles FAQ © 2021—2026. All Rights Reserved. Privacy Policy | Terms of Services Download Bee Network APP and start the web3 journey The world's largest Web3 portal Partners CoinCarp Binance CoinMarketCap CoinGecko Coinlive Armors White Paper Roles FAQ © 2021—2026. All Rights Reserved. Privacy Policy | Terms of Services Search SearchInSiteOnChainSocialNews Hot to you: Airdrop Hunters Data Analysis Crypto Celebrities Trap Detector English 繁體中文 简体中文 日本語 Tiếng Việt العربية 한국어 Bahasa Indonesia हिन्दी اردو Русский English
智能索引记录
-
2026-03-02 10:08:05
综合导航
成功
标题:ISG job portal - FAQ
简介:Here you will find frequently asked questions and answers ab
-
2026-03-02 23:04:03
综合导航
成功
标题:æ¸
æçæ¼é³_æ¸
æçææ_æ¸
æçç¹ä½_è¯ç»ç½
简介:è¯ç»ç½æ¸ æé¢é,ä»ç»æ¸ æ,æ¸ æçæ¼é³,æ¸ ææ¯
-
2026-03-03 01:45:42
综合导航
成功
标题:Letter to the Editor: Responding to OEHHA on PFAs Law.com
简介:According to the authors, the Office of Environmental Health
-
2026-03-03 06:55:35
综合导航
成功
标题:Take Back Your Workspace: The KonMari Method at Work Inspiring Workspaces by BOS
简介:Since the KonMari method is so popular in the home, it makes
-
2026-03-02 20:16:32
综合导航
成功
标题:Secure Investment Management Ranks #158 on the 2025 Inc. 5000 List of America’s Fastest-Growing Private Companies — A 78% Leap in National Ranking - PR.com
简介:Secure Investment Management ranked #158 on the 2025 Inc. 50
-
2026-03-02 22:05:48
综合导航
成功
标题:거산고구마 상품 후기 달콤하고 쫀득해요
简介:거산고구마 자연 그대로의 달콤함, 신선한 맛으로 고객 만족이 높아요
-
2026-03-03 04:54:12
综合导航
成功
标题:Your Timetable RGU
简介:To access your individual timetable, choose one of these eas
-
2026-03-03 05:16:29
新闻资讯
成功
标题:采用第二代AMD EPYC的PowerEdge服务器在戴尔科技峰会, 站长资讯平台
简介:在近日于北京举行的2019戴尔科技峰会上,戴尔科技集团与AMD公司携手,正式推出了针对第二代AMD EPYC 处理器而设
-
2026-03-02 22:10:15
视频影音
成功
标题:这可太心累了,什么时候到头,赶紧找充电桩!_网易视频
简介:这可太心累了,什么时候到头,赶紧找充电桩!
-
2026-03-03 07:14:52
游戏娱乐
成功
标题:欢乐麻将2.5,欢乐麻将2.5小游戏,4399小游戏 www.4399.com
简介:欢乐麻将2.5在线玩,欢乐麻将2.5下载, 欢乐麻将2.5攻略秘籍.更多欢乐麻将2.5游戏尽在4399小游戏,好玩记得告
-
2026-03-02 22:01:39
综合导航
成功
标题:Meet Our Specialists PwC Canada
简介:Data and Analytics is at the core of everything we do as PwC
-
2026-03-02 20:22:08
综合导航
成功
标题:Financing Express of the Week 11 projects received investment, with a total disclosed financing amount of US$98.5 mill Bee Network
简介:According to incomplete statistics from Odaily Planet Daily,
-
2026-03-02 21:45:22
旅游出行
成功
标题:宁波一旅游团所乘邮轮滞留迪拜,船上有约200名中国游客 巴林 宁波市 邮轮公司 外国游客_网易订阅
简介:宁波一旅游团所乘邮轮滞留迪拜,船上有约200名中国游客,迪拜,巴林,旅游团,宁波市,中国游客,邮轮公司,外国游客
-
2026-03-03 05:21:18
综合导航
成功
标题:XS: Forex Trading & CFDs Broker Online FX Trading Platform
简介:Discover the leading online forex trading platform at XS. Tr
-
2026-03-02 20:35:27
综合导航
成功
标题:éèçæ¼é³_éèçææ_éèçç¹ä½_è¯ç»ç½
简介:è¯ç»ç½éèé¢é,ä»ç»éè,éèçæ¼é³,éèæ¯
-
2026-03-03 02:28:24
综合导航
成功
标题:九酷音乐网 流行歌曲 经典歌曲 mp3音乐歌曲免费下载在线听
简介:音乐-歌曲.九酷音乐网是专业的在线音乐试听mp3下载网站.收录了网上最新歌曲和流行音乐,网络歌曲,好听的歌,抖音热门歌曲
-
2026-03-02 22:08:45
博客创作
成功
标题:乱世:繁衍子嗣,从娶妻开始争霸txt下载_乱世:繁衍子嗣,从娶妻开始争霸全本 全文 全集 完本_最新版小说电子书免费下载_新笔趣阁(56xu.com)
简介:新笔趣阁提供乱世:繁衍子嗣,从娶妻开始争霸txt下载,青水师兄创作的玄幻奇幻小说乱世:繁衍子嗣,从娶妻开始争霸txt免费
-
2026-03-02 15:12:14
综合导航
成功
标题:DeriW: The First Zero-Fee Contract ExchangeRecommended Articles Bee Network
简介:去中心化衍生品交易所(Perp DEX)已成为了当前市场上最热的赛道之一。 一方面,过去几年间 Perp DEX 正不断
-
2026-03-02 21:55:38
综合导航
成功
标题:WTB igniter/coil module [Archive] - Toyota MR2 Message Board
简介:I have a 1991 NA MR2, I need the coil module and igniter mod
-
2026-03-02 22:54:49
综合导航
成功
标题:Celestia Bee Network
简介:Jaringan blockchain modular pertama yang memudahkan siapa pu
-
2026-03-02 13:59:00
教育培训
成功
标题:[优]五年级的作文8篇
简介:在平凡的学习、工作、生活中,许多人都写过作文吧,作文是人们把记忆中所存储的有关知识、经验和思想用书面形式表达出来的记叙方
-
2026-03-02 21:53:33
综合导航
成功
标题:笔记本触摸板失灵?Synaptics驱动下载与安装方法-驱动人生
简介:Synaptics触摸板驱动怎么下载?本文详细介绍Synaptics触摸板驱动的作用,并提供多种下载安装方法,包括使用驱
-
2026-03-03 06:59:31
运动健身
成功
标题:腊梅初中作文6篇
简介:无论是在学校还是在社会中,大家总少不了接触作文吧,写作文可以锻炼我们的独处习惯,让自己的心静下来,思考自己未来的方向。相
-
2026-03-02 10:21:17
综合导航
成功
标题:Labels, Stickers, & Decals PIP - PIP East Longmeadow, MA
简介:Get your decals and stickers created and printed quickly for
-
2026-03-02 22:00:26
新闻资讯
成功
标题:数据科学公司如何使用漫画来可视化数据, 站长资讯平台
简介:by Cate Lawrence 来源:DZone 即便你是世界上最好的作者,但是如果人们不阅读你的作品,没人会收到
-
2026-03-02 21:58:53
综合导航
成功
标题:韫是五行中的哪个,韫是五行中的哪个部位? - 吉吉算命网
简介:【导读】吉吉算命网分享“韫是五行中的哪个,韫是五行中的哪个部位?”的解读,解惑找吉吉算命网,韫是五行中的哪个部位,韫是五
-
2026-03-03 07:11:50
数码科技
成功
标题:电脑右下角出现激活windows 三种快速解决方法指南-驱动人生
简介:电脑屏幕右下角不断显示“激活 Windows,请转到设置以激活 Windows”的水印。本文介绍三种快速且有效的解决方法
-
2026-03-03 05:23:06
图片素材
成功
标题:好看的作文550字 描写好看的作文 关于好看的作文-作文网
简介:作文网精选关于好看的550字作文,包含好看的作文素材,关于好看的作文题目,以好看为话题的550字作文大全,作文网原创名师
-
2026-03-02 22:00:58
综合导航
成功
标题:Height Adjustable Options at Work Inspiring Workspaces by BOS
简介:Thinking of adding height adjustable surfaces to your worksp
-
2026-03-02 23:00:19
教育培训
成功
标题:写爷爷的作文汇总九篇
简介:在平凡的学习、工作、生活中,大家都写过作文,肯定对各类作文都很熟悉吧,写作文是培养人们的观察力、联想力、想象力、思考力和