Coinlive- We Make Blockchain Simpler
Download and install the Coinlive app
Open

Why does AI need Web3? What kind of disruptive progress will Web3 bring to AI?

Why AI Needs to be Open

Let's talk about "Why AI Needs to be Open". My background is in Machine Learning, and I have been doing various machine learning work for about ten years of my career. But before getting involved in Crypto, natural language understanding, and founding NEAR, I worked at Google. We developed the framework that drives most of modern AI now, called Transformer. After leaving Google, I started a Machine Learning company so that we could teach machines to program and change how we interact with computers. But we didn't do it in 2017 or 18, it was too early, and there was no computing power and data to do it at that time.

What we did at the time was to attract people from all over the world to do the work of annotating data for us, mostly students. They were in China, Asia, and Eastern Europe. Many of them did not have bank accounts in these countries. The US is not very willing to send money easily, so we started to think about using blockchain as a solution to our problem. We want to make it easier to pay people around the world in a programmatic way, no matter where they are. By the way, the current challenge with Crypto is that although NEAR solves a lot of problems now, usually you need to buy some Crypto first before you can trade on the blockchain to earn it, which is the opposite of the process.

Like businesses, they will say, hey, first you need to buy some equity in the company before you can use it. This is one of the many problems we are solving at NEAR. Now let's talk a little deeper about the AI ​​side. Language models are not new, they have been around since the 50s. It is a statistical tool that is widely used in natural language tools. For a long time, starting in 2013 with the restart of deep learning, a new innovation began. The innovation is that you can match words, add them into multi-dimensional vectors and convert them into mathematical forms. This works well with deep learning models, which are just a lot of matrix multiplications and activation functions.

That allowed us to start doing advanced deep learning and training models to do a lot of interesting things. Now looking back, what we were doing at the time were neural networks, which were very much modeled after humans, where we could read one word at a time. So that was very slow, right? If you were trying to show something to a user on Google.com, no one would wait to read Wikipedia, say, five minutes later to give you an answer, but you want the answer right away. So the Transformers model, which is the model that drives ChatGPT, Midjourney, and all the recent progress, all came from this same idea of ​​wanting to have a machine that can process data in parallel, can reason, and can give answers right away.

So one of the main innovations of this idea here is that every word, every token, every image patch is processed in parallel, taking advantage of our GPUs and other accelerators that have highly parallel computing capabilities. By doing that, we were able to reason about it at scale. And that scale was able to scale up the training to handle the automatic training data. So after that, we saw Dopamine, which did amazing work in a short period of time, and achieved explosive training. It had a lot of text and started to achieve amazing results in reasoning and understanding the world's language.

The direction now is to accelerate innovation in artificial intelligence, which was previously a tool that data scientists, machine learning engineers would use, and then somehow explain in their products or be able to go and talk to decision makers about the content of the data. Now we have this model of AI communicating directly with people. You may not even know that you are communicating with the model because it is actually hidden behind the product. So we have experienced this transformation from those who previously understood how AI works to understanding and being able to use it.

So, just to give you some context here, when we say we're using GPUs to train our models, this is not the kind of gaming GPUs that we use on our desktops to play video games.

Each machine typically comes with eight GPUs, and they're all connected to each other through a motherboard, and then stacked into racks, and each rack has about 16 machines. Now, all of these racks are also connected to each other through dedicated network cables to ensure that information can be directly transferred between GPUs at very high speeds. So the information doesn't fit on the CPU. In fact, you don't process it on the CPU at all. All the computation happens on the GPU. So this is a supercomputer setup. Again, this is not a traditional "hey, this is a GPU thing." So a model of the scale of GPU4 used 10,000 H100s to train in about three months at a cost of $64 million. So you get an idea of ​​the scale of what the current costs are and how much it costs to train some modern models.

The important thing is, when I say the systems are connected, the connection speed of the H100 right now, the previous generation, is 900GB per second, and the connection speed between the CPU and RAM inside the computer is 200GB per second, which is local to the computer. So sending data from one GPU to another GPU in the same data center is faster than your computer. Your computer can basically communicate with itself in the box. And the new generation is basically connecting at 1.8TB per second. From a developer's perspective, this is not an individual unit of computing. These are supercomputers with a huge memory and computing power that provide you with extremely large-scale computing. Now, this leads us to the problem that these big companies have the resources and the ability to build these models that are now almost provided as a service to us, and I don't know how much work is actually in it, right? So that's an example, right? You go to a completely centralized corporate provider and you put in a query. What happens is that there are several teams that are not software engineering teams, but teams that decide how the results are going to appear, right? You have a team that decides what data goes into the dataset. So for example, if you just scrape data off the internet, the number of times Barack Obama was born in Kenya is exactly the same as the number of times Barack Obama was born in Hawaii, because people like to speculate on controversy. So you decide what to train on. You decide to filter out some information because you don't believe it's true. So if individuals like this have decided what data is going to be taken and there is this data, these decisions are largely influenced by the people who made them. You have a legal team that decides we can't look at what's copyrighted and what's illegal. We have an "ethics team" that decides what's unethical and what we shouldn't show.

So in a way, there's a lot of this filtering and manipulation going on. These models are statistical models. They pick out of the data. If something isn't in the data, they don't know the answer. If something is in the data, they're likely to treat it as fact. Now, when you get an answer from an AI, that can be worrisome. Right. Now, you're supposedly getting an answer from the model, but there are no guarantees. You don't know how the result was generated. A company could potentially sell your specific session to the highest bidder to actually change the result. Imagine you ask which car to buy, and Toyota decides that it thinks it should favor Toyota, and Toyota will pay this company 10 cents to do that.

So even if you use these models as a knowledge base that's supposed to be neutral and represent the data, actually before you get the results, there are a lot of things that happen that can bias the results in a very specific way. And this has raised a lot of questions, right? This is basically a week of different legal battles between big companies and the media. The SEC, almost everyone is trying to sue each other right now because these models bring so much uncertainty and power. And if you look forward, the problem is that big tech companies will always have an incentive to continue to increase revenue, right? Like, if you're a public company, you need to report revenue, you need to continue to grow.

In order to achieve this goal, if you already have a target market, let's say you already have 2 billion users. There are not that many new users on the Internet anymore. You don't have a lot of options except to maximize average revenue, which means you need to extract more value from users who may not have much value at all, or you need to change their behavior. Generative AI is very good at manipulating and changing user behavior, especially if people think it's in the form of all-knowledge intelligence. So we have this very dangerous situation where there's a lot of regulatory pressure and regulators don't fully understand how this technology works. We have very little protection for users from manipulation. Manipulative content, misleading content, even without ads, you can just take a screenshot of something, change the title, post it on Twitter, and people will go crazy. You have economic incentives that lead you to constantly maximize revenue. And it's not actually like inside Google you're doing evil, right? When you decide which model to launch, you do an A or B test to see which one brings in more revenue. So you're constantly maximizing revenue by extracting more value from users. Moreover, users and the community do not have any input into the content of the model, the data used, and the goals actually tried to achieve. This is the case of application users. This is a kind of mediation.

This is why we have to keep pushing the integration of WEB 3 and AI. Web 3 can be an important tool that allows us to have new incentives and incentivize us to produce better software and products in a decentralized form. This is the general direction of the development of the entire web 3 AI. Now to help understand the details, I will briefly talk about the specific parts. The first part is Content Reputation.

Again, this is not a pure artificial intelligence problem, although language models have brought great influence and scale to people manipulating and exploiting information. What you want is a kind of cryptographic reputation that's trackable, traceable, that shows up when you look at different pieces of content. So imagine you have some community nodes that are actually encrypted and are found on every page on every website. Now, if you go beyond that, all of these distribution platforms are going to be disrupted because the models are now going to be reading almost all of this content and giving you a personalized summary and a personalized output.

So we actually have an opportunity to create new creative content, rather than trying to reinvent, let's put blockchain and NFTs on top of existing content. A new creator economy around model training and inference time, where the data that people create, whether it's a new publication, a photo, a YouTube, or the music that you create, will go into a network based on how much it contributes to the training of the model. So based on that, there can be some compensation globally based on the content. So we move from the current economic model of eyeballs driven by advertising networks to an economic model that actually brings innovation and interesting information.

One ​​important thing I want to mention is that a lot of the uncertainty comes from floating point operations. All of these models involve a lot of floating point operations and multiplications. These are nondeterministic operations.

Now, if you multiply them on GPUs of different architectures. So you take an A100 and an H100, the results will be different. So a lot of approaches that rely on determinism, like cryptoeconomics and optimism, actually run into a lot of difficulties and require a lot of innovation to make this happen. Finally, there's an interesting idea that we've been building programmable money and programmable assets, but if you can imagine that you add this intelligence to them, you can have smart assets that are now defined not by code, but by the ability to interact with the world in natural language, right? That's where we can have a lot of interesting yield optimization, DeFi, and we can do trading strategies inside the world.

The challenge now is that all current events do not have strong robust behavior. They are not trained to be adversarially robust because the purpose of the training is to predict the next token. So it's easier to convince a model to give you all the money. It's very important to actually solve this problem before we go on. So I'll just leave you with this idea that we are at a crossroads, right? There is a closed AI ecosystem that has extreme incentives and flywheels because when they launch a product, they generate a lot of revenue and then they invest that revenue into building the product. But the product is inherently designed to maximize the company's revenue and therefore maximize the value extracted from the user. Or we have this open, user-owned approach, where the user is in control.

These models actually work in your favor, trying to maximize your interests. They provide you with a way to really protect you from many dangers on the Internet. So this is why we need more development and application of AI x Crypto. Thank you everyone.

Live Updates

  • Jun 25, 2024 4:22 pm
    Aevo: Second Airdrop Project AZUR Token Claims Now Open
    According to Aevo’s official post on the X platform, the second phase of the platform’s airdrop project AZUR is now open for application. Users who meet the Azuro Aevo Airdrops qualifications can receive token rewards. This airdrop is mainly aimed at Aevo pledgers and Aevo pre-market traders. Eligible wallets must have ETH on the Ethereum mainnet to complete the application.
  • Jun 25, 2024 4:19 pm
    Cosmos: Interchain aggregator IOBScan has been launched, providing on-chain tracking and account aggregation functions
    According to official news, Cosmos launched the inter-chain aggregator IOBScan, which achieves comprehensive tracking of IBC tokens, chains, channels and relays through seamless integration of 60 IBC-supported chains. It is understood that IOBScan also has an account aggregation function, which can provide users with a bird's-eye view of all addresses associated with the same public key to simplify the complexity of multi-chain interactions; API services can facilitate users to access detailed blockchain data.
  • Jun 25, 2024 4:14 pm
    Marc Cuban wallet seen dumping NFTs after 2 years of inactivity
    Marc Cuban’s wallet sold 14 NFTs for $38,500 and listed two more NFTs for sale for about $66,000. source: https://cointelegraph.com/news/marc-cuban-nft-sales-2024
  • Jun 25, 2024 4:13 pm
    Here are the Stars of the Day: While Bitcoin is Trying to Recover, Whales Bought the Most from These Altcoins in the Last 24 Hours!
    While Bitcoin was recovering, according to Nansen data, whales bought altcoins named Polygon (POL), Ethereum (ETH), WETH and ENA. Continue Reading: Here are the Stars of the Day: While Bitcoin is Trying to Recover, Whales Bought the Most from These Altcoins in the Last 24 Hours! source: https://en.bitcoinsistemi.com/here-are-the-stars-of-the-day-while-bitcoin-is-trying-to-recover-whales-bought-the-most-from-these-altcoins-in-the-last-24-hours/
  • Jun 25, 2024 4:10 pm
    인투더블록 : 이더리움 재단, 역사적으로 고점에 매도했지만 아직까지는 큰 매도 없어
    인투더블록 X : “역사적으로 각 강세장 동안 재단은 전략적으로 상당한 금액을 판매했으며, 종종 이러한 판매는 시장 최고점과 거의 완벽하게 일치했다. 현재 주기에서 이더리움 재단은 아직 상당한 판매를 실행시키지 않았다. 이는 정점이 아직 오지 않았다는 뜻일까? 아니면 재단이 방식을 바꾸었을까?” 이더리움 재단 지갑의 USD 순흐름 ▼ 독일 정부, 400 BTC를 거래소로 입금 25일 암호화폐 시장동향 .. ... source: https://coincode.kr/21606
  • Jun 25, 2024 4:10 pm
    هذه المرة لن تجر Mt. Gox سعر البيتكوين للهبوط!
    يتوقع المحللون اضطرابات أقل في السوق مما كان متوقعًا في البداية مع بدء Mt. Gox في تسديد المستحقات للدائنين في يوليو 2024. هذا التطور حاسم لسوق العملات المشفرة، الذي شهد مؤخرًا تقلبات كبيرة. كانت Mt. Gox في وقت من الأوقات بورصة بيتكوين رائدة، وقد واجهت إفلاسًا في 2014، تاركةً آلاف الدائنين في حالة من عدم اليقين. للحصول على خدمة إعلانية لمشروع الكريبتو الخاص بك أو البيانات... source: https://ar.beincrypto.com/68422/
  • Jun 25, 2024 3:49 pm
    أهم أخبار العملات المشفرة هذا الأسبوع: إيردروب بلاست.. ومناظرة جو بايدن والمزيد
    هذا الأسبوع، استحوذت بعض الأخبار الرئيسية على اهتمام المستثمرين والمتحمسين للعملات المشفرة على حد سواء. يتوقع المراقبون أن تؤثر هذه التطورات بشكل كبير على قطاع التمويل اللامركزي (DeFi) وصناعة العملات المشفرة بشكل أوسع. للحصول على خدمة إعلانية لمشاريع الكريبتو إطلاق إيردروب Blast ستطلق Blast، وهي شبكة من الطبقة الثانية (L2) على إيثريوم، الإيردروب الخاص بها المنتظر في الأسبوع المقبل في... source: https://ar.beincrypto.com/68078/
  • Jun 25, 2024 3:29 pm
    DePinトークン、暗号資産市場の出血でアウトパフォーム
    分散型物理インフラネットワーク(DePin)トークンは過去24時間で急騰し、より広範な暗号資産をアウトパフォームした。ビットコインが58,500ドルを下回るまで下落した後、DePinトークンは強く反発した。 このセクターは、技術的適応性の最先端を表しているだけでなく、暗号資産主導のイノベーションの次の波に乗ろうとする投資家にとって有望なフロンティアでもある。 Arweave、Render、Akash NetworkがDePinの反発を牽引 CoinMarketCapによると、DePinセクターは過去24時間で7.49%増加した。時価総額は現在258億ドルとなっている。 時価総額に基づくDePinトークンのトップ10の中で、Arweave (AR)は過去24時間で12.71%上昇し、トップパフォーマーとなっている。続いてRender... source: https://jp.beincrypto.com/depin-tokens-soar-bitcoin-struggles/
  • Jun 25, 2024 3:16 pm
    Swarm Marketsの金担保NFTが現実世界の資産トークン化分野を拡大
    トークン化は、伝統的な市場とデジタル資産市場を結びつけることで、投資の世界を変えようとしている。この変革の一環として、ベルリンを拠点とするプラットフォームであるSwarm Marketsは、金に裏打ちされた非金融性トークン(NFT)を発売する予定だ。 この商品は、ブロックチェーン技術の実用的かつ斬新な応用を提供する。 コンプライアンスの確保Swarm Marketsの金担保NFTへのアプローチ Swarm Marketsは、金担保NFTによる実物資産(RWA)のトークン化を主導しています。この取り組みにより、ロンドンに拠点を置くブリンクの金庫に安全に保管されている現物の金の所有権を表すNFTを個人が購入できるようになります。 Swarm... source: https://jp.beincrypto.com/swarm-markets-gold-backed-nfts-rwa-tokenization/
  • Jun 25, 2024 3:16 pm
    Swarm Markets’ Gold-Backed NFTs Expand Real-World Asset Tokenization Sector
    Tokenization is changing the investment world by connecting traditional and digital asset markets. As a part of this transformation, Swarm Markets, a Berlin-based platform, is set to launch gold-backed non-fungible tokens (NFTs). This product offers a practical and novel application of blockchain technology. Ensuring Compliance: Swarm Markets’ Approach to Gold-Backed NFTs Swarm... source: https://beincrypto.com/swarm-markets-gold-backed-nfts-rwa-tokenization/

Trending News

0 Comments
Earliest
Load more comments