{"id":23018,"date":"2024-01-23T05:49:43","date_gmt":"2024-01-23T05:49:43","guid":{"rendered":"https:\/\/web3unplugged.io\/blog\/?p=23018"},"modified":"2024-01-23T05:49:45","modified_gmt":"2024-01-23T05:49:45","slug":"demystifying-ai-the-probability-theory-behind-llms-like-openais-chatgpt","status":"publish","type":"post","link":"https:\/\/web3unplugged.io\/blog\/demystifying-ai-the-probability-theory-behind-llms-like-openais-chatgpt\/","title":{"rendered":"Demystifying AI: The Probability Theory Behind LLMs Like OpenAI\u2019s ChatGPT"},"content":{"rendered":"\n<p>When a paradigm shift occurs, it is not always obvious to those affected by it.&nbsp;<\/p>\n\n\n\n<p>But there is no \u201ceye of the storm\u201d equivalent when it comes to generative artificial intelligence (AI).&nbsp;<\/p>\n\n\n\n<p>The technology is&nbsp;here.&nbsp;There are already various&nbsp;commercial products&nbsp;available for&nbsp;deployment, and organizations that can effectively leverage it in support of their&nbsp;business goals&nbsp;are likely to outperform their peers that fail to adopt the innovation.&nbsp;<\/p>\n\n\n\n<p>Still, as with many innovations, uncertainty and institutional inertia reign supreme \u2014 which is why understanding how the large language models (LLMs) powering AI work is critical to not just piercing the black box of the technology\u2019s supposed inscrutability, but also to applying AI tools correctly within an enterprise setting.&nbsp;<\/p>\n\n\n\n<p>The most important thing to understand about the foundational models powering today\u2019s AI interfaces and giving them their ability to generate responses is the simple fact that LLMs, like Google\u2019s Bard, Anthropic\u2019s Claude, OpenAI\u2019s ChatGPT and others, are just adding one word at a time.&nbsp;<\/p>\n\n\n\n<p>Underneath the layers of sophisticated algorithmic calculations, that\u2019s all there is to it.&nbsp;<\/p>\n\n\n\n<p>That\u2019s because at a fundamental level, generative AI models are built to generate reasonable continuations of text by drawing from a ranked list of words, each given different weighted probabilities based on the data set the model was trained on.&nbsp;<\/p>\n\n\n\n<p>How AI Works<br>While news of AI that can surpass human intelligence are helping fuel the hype of the technology, the reality is far more driven by math than it is by myth.&nbsp;<\/p>\n\n\n\n<p>\u201cIt is important for everyone to understand that AI&nbsp;learns from data&nbsp;\u2026 at the end of the day [AI] is merely probabilistics and statistics,\u201d&nbsp;Akli Adjaoute, AI pioneer and founder and general partner at venture capital fund&nbsp;Exponion, told PYMNTS in November.&nbsp;<\/p>\n\n\n\n<p>But where do the probabilities that determine an AI systems\u2019 output originate from?&nbsp;<\/p>\n\n\n\n<p>The answer lies within the&nbsp; AI model\u2019s training data. Peeking into the inner workings of an AI model reveals that it is not only the next reasonable word that is being identified, weighted, then generated, but that this process occurs on a letter by letter basis, as AI models break apart words into more&nbsp;manageable tokens.&nbsp;<\/p>\n\n\n\n<p>That is a big part of why&nbsp;prompt engineering&nbsp;for AI models is an emerging skillset. After all, different prompts produce different outputs based on the probabilities inherent to each reasonable continuation, meaning that to get the best output, you need to have a clear idea of where to point the provided input or query.&nbsp;<\/p>\n\n\n\n<p>It also means that the data informing the weight given to each probabilistic outcome must be&nbsp;relevant&nbsp;to the query. The more relevant, the better.&nbsp;<\/p>\n\n\n\n<p>See also:&nbsp;Tailoring AI Solutions by Industry Key to Scalability<\/p>\n\n\n\n<p>Making AI Work for You<br>While PYMNTS Intelligence has found that more than eight in 10 business leaders (84%) believe generative AI\u00a0will positively impact\u00a0the workforce, generative AI systems are only as good as the data they\u2019re trained on. That\u2019s why the largest AI players are in an arms race to\u00a0acquire the best training data sets.<\/p>\n\n\n\n<p>\u201cThere\u2019s a long way to go before there\u2019s a&nbsp;futuristic version of AI&nbsp;where machines think and make decisions. \u2026 Humans will be around for quite a while,\u201d&nbsp;&nbsp;Tony Wimmer, head of data and analytics at&nbsp;J.P. Morgan Payments, told PYMNTS in March. \u201cAnd the more that we can write software that has payments data at the heart of it to help humans, the better payments will get.\u201d<\/p>\n\n\n\n<p>That\u2019s why, to train an AI model to perform to the necessary standard, many enterprises are relying on&nbsp;their own internal data&nbsp;to avoid compromising model outputs. By creating vertically specialized LLMs trained for industry use cases, organizations can deploy AI systems that are able to find the signal within the noise, as well as to be further fine-tuned to business-specific goals with real-time data.&nbsp;<\/p>\n\n\n\n<p>As&nbsp;Akli Adjaoute&nbsp;told PYMNTS back in November, \u201cif you go into a field where the data is real, particularly in the&nbsp;payments industry, whether it\u2019s credit risk, whether it\u2019s delinquency, whether it\u2019s AML [anti-money laundering], whether it\u2019s fraud prevention, anything that touches payments \u2026 AI can bring a lot of benefit.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"<p>When a paradigm shift occurs, it is not always obvious to those affected by it.&nbsp; But there is no \u201ceye of the storm\u201d equivalent when it comes to generative artificial intelligence (AI).&nbsp; The technology is&nbsp;here.&nbsp;There are already various&nbsp;commercial products&nbsp;available for&nbsp;deployment, and organizations that can effectively leverage it in support of their&nbsp;business goals&nbsp;are likely to outperform [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":23020,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_seopress_robots_primary_cat":"none","_seopress_titles_title":"","_seopress_titles_desc":"","_seopress_robots_index":"","footnotes":""},"categories":[2],"tags":[],"class_list":["post-23018","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news"],"rttpg_featured_image_url":{"full":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"landscape":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"portraits":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"thumbnail":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18-150x150.jpg",150,150,true],"medium":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18-300x180.jpg",300,180,true],"large":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"1536x1536":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"2048x2048":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"post-thumbnail":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",620,372,false],"graptor-sq-xs":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2024\/01\/Untitled-18.jpg",100,60,false]},"rttpg_author":{"display_name":"Admin CG","author_link":"https:\/\/web3unplugged.io\/blog\/author\/admin-cg\/"},"rttpg_comment":0,"rttpg_category":"<a href=\"https:\/\/web3unplugged.io\/blog\/category\/news\/\" rel=\"category tag\">news<\/a>","rttpg_excerpt":"When a paradigm shift occurs, it is not always obvious to those affected by it.&nbsp; But there is no \u201ceye of the storm\u201d equivalent when it comes to generative artificial intelligence (AI).&nbsp; The technology is&nbsp;here.&nbsp;There are already various&nbsp;commercial products&nbsp;available for&nbsp;deployment, and organizations that can effectively leverage it in support of their&nbsp;business goals&nbsp;are likely to outperform&hellip;","_links":{"self":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts\/23018","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/comments?post=23018"}],"version-history":[{"count":1,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts\/23018\/revisions"}],"predecessor-version":[{"id":23021,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts\/23018\/revisions\/23021"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/media\/23020"}],"wp:attachment":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/media?parent=23018"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/categories?post=23018"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/tags?post=23018"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}