{"id":8777,"date":"2024-10-02T08:58:15","date_gmt":"2024-10-02T08:58:15","guid":{"rendered":"https:\/\/wallstwarroom.com\/uncategorized\/the-future-of-ai-how-edge-computing-is-reshaping-the-landscape-from-cloud-to-handheld-devices\/"},"modified":"2024-10-02T08:58:15","modified_gmt":"2024-10-02T08:58:15","slug":"the-future-of-ai-how-edge-computing-is-reshaping-the-landscape-from-cloud-to-handheld-devices","status":"publish","type":"post","link":"https:\/\/wallstwarroom.com\/h\/tech-stocks\/the-future-of-ai-how-edge-computing-is-reshaping-the-landscape-from-cloud-to-handheld-devices\/","title":{"rendered":"The Future of AI: How Edge Computing is Reshaping the Landscape from Cloud to Handheld Devices"},"content":{"rendered":"<h1>AI&#8217;s Next Feat: The Descent from the Cloud<\/h1>\n<p>It&#8217;s been two years since <strong>ChatGPT<\/strong> made its public debut, igniting a wave of investment in generative artificial intelligence (AI). This frenzy has driven up valuations for startups like <strong>OpenAI<\/strong>, the inventor of the chatbot, and for major technology companies whose cloud computing platforms train and host the AI models that power these applications. However, the current boom is starting to show signs of strain. The next phase of AI growth may be in the palm of users\u2019 hands, as innovations in edge computing come to the forefront.<\/p>\n<h2>Current AI Landscape: The Role of the Cloud<\/h2>\n<p>Generative AI, which revolves around models that create new content based on their training data, is largely cloud-dependent at present. For instance, OpenAI utilizes <strong>Microsoft Azure<\/strong> to train and operate its large language models (LLMs). Users from across the globe can access ChatGPT through Azure\u2019s extensive network of data centers. However, as these models grow in size and complexity, so too does the underlying infrastructure required to train them and respond to user inquiries.<\/p>\n<p>The result? A frenzied race to develop larger and more powerful data centers. OpenAI and Microsoft are currently in discussions regarding a massive data center project scheduled for a 2028 launch, with projected costs hitting an astonishing <strong>$100 billion<\/strong>, according to reports from <strong>The Information<\/strong>. Overall, tech giants such as <strong>Google<\/strong> (owner of Alphabet), Microsoft, and <strong>Meta Platforms<\/strong> (the company behind Instagram and Facebook) are expected to collectively spend around <strong>$160 billion<\/strong> on capital expenditures next year, a staggering <strong>75% increase<\/strong> compared to 2022. Most of these expenses will go toward securing <strong>Nvidia&#8217;s<\/strong> highly sought-after $25,000 graphic processor units (GPU) and the necessary infrastructure for model training.<\/p>\n<h2>Technological Hurdles: Challenges on the Horizon<\/h2>\n<p>The largest hurdle the industry faces is technological. Today&#8217;s smartphones and devices lack the computing power, energy, and memory bandwidth necessary to run an expansive model like OpenAI\u2019s <strong>GPT-4<\/strong>, which contains approximately <strong>1.8 trillion parameters<\/strong>. Even smaller models like Facebook&#8217;s LLAMA, containing 7 billion parameters, would demand an additional <strong>14 GB<\/strong> of temporary storage\u2014an impractical feat for current smartphones. For example, <strong>Apple&#8217;s iPhone 16<\/strong> only offers <strong>8 GB<\/strong> of RAM.<\/p>\n<h2>Optimism on the Horizon: A Shift Toward Smaller Models<\/h2>\n<p>Despite these challenges, there\u2019s room for optimism. Companies and developers are increasingly turning to streamlined models tailored for specific tasks. These smaller models require less data and effort to train, and they are often open-source and freely accessible. Google&#8217;s newly introduced &#8220;lightweight&#8221; model, <strong>Gemma<\/strong>, exemplifies this trend with only <strong>2 billion parameters<\/strong>. Their specialized nature frequently allows them to outperform larger, more generalized models while exhibiting fewer errors.<\/p>\n<p>Moreover, many everyday uses of AI, including photo-editing tools and personal assistants, likely won\u2019t necessitate the expense of extensive models. Several smartphones already incorporate live translation and real-time transcription capabilities. Thus, it\u2019s logical for cloud providers to transition basic AI functionalities to edge devices, reserving dense data centers for more complex tasks.<\/p>\n<h2>The Rise of Advanced Semiconductors<\/h2>\n<p>Additionally, advancements in semiconductor technology are propelling the capabilities of devices. Research firm <strong>Yole Group<\/strong> estimates that the proportion of smartphones that can support an LLM with 7 billion parameters is projected to increase to <strong>11% this year<\/strong>, up from <strong>8%<\/strong> last year. Leading chip manufacturers, including Taiwan&#8217;s <strong>TSMC<\/strong> and South Korea&#8217;s <strong>Samsung Electronics<\/strong> and <strong>SK Hynix<\/strong>, are developing cutting-edge techniques such as advanced chip packaging, which involves stacking multiple chips into a single &#8220;chiplet.&#8221; This innovation enables them to create more powerful processors by consolidating more transistors without reducing chip circuitry size.<\/p>\n<h2>Investment Opportunities: The Edge AI Market<\/h2>\n<p>For investors, the burgeoning field of edge AI holds the promise of generating new winners. Up until now, market assumptions have centered on large tech firms, deep-pocketed giants, Nvidia, and a select few startups capturing the majority of AI&#8217;s economic boosts. However, the introduction of AI-enhanced tools has the potential to drive consumers toward upgrading to sophisticated smartphones and personal computers. <strong>UBS analysts<\/strong> predict that sales in these markets will exceed <strong>$700 billion<\/strong> by 2027, reflecting a <strong>14% increase<\/strong> from this year.<\/p>\n<p>Brands ranging from <strong>Apple<\/strong> to <strong>Lenovo<\/strong>, along with their respective suppliers, stand to benefit from this trend. While Nvidia&#8217;s sophisticated GPUs are likely to continue leading the market, other chip manufacturers, such as <strong>Qualcomm<\/strong> and <strong>MediaTek<\/strong>, are also poised to gain. MediaTek plans to unveil its latest chipset capable of supporting large models next month, predicting a <strong>50% growth<\/strong> in revenue from its flagship mobile products this year.<\/p>\n<h2>Conclusion: The Next Big Thing in AI<\/h2>\n<p>The forthcoming success of edge AI will hinge on developers creating compelling applications that users find valuable. Should this reality transpire, the next significant evolution in AI may lie within smaller models and devices, reshaping the landscape and impacting how consumers interact with technology.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>AI&#8217;s Next Feat: The Descent from the Cloud It&#8217;s been two years since ChatGPT made its public debut, igniting a wave of investment in generative artificial intelligence (AI). This frenzy has driven up valuations for startups like OpenAI, the inventor of the chatbot, and for major technology companies whose cloud computing platforms train and host&#8230;<\/p>\n","protected":false},"author":32,"featured_media":8776,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_eb_attr":"","footnotes":""},"categories":[683],"tags":[],"class_list":["post-8777","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-tech-stocks"],"_links":{"self":[{"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/posts\/8777","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/users\/32"}],"replies":[{"embeddable":true,"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/comments?post=8777"}],"version-history":[{"count":0,"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/posts\/8777\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/media\/8776"}],"wp:attachment":[{"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/media?parent=8777"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/categories?post=8777"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/wallstwarroom.com\/h\/wp-json\/wp\/v2\/tags?post=8777"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}