Path: ...!news.roellig-ltd.de!open-news-network.org!weretis.net!feeder8.news.weretis.net!reader5.news.weretis.net!news.solani.org!.POSTED!not-for-mail From: Mild Shock Newsgroups: sci.math Subject: =?UTF-8?Q?Re:_LLM_versus_CYC_=28Was:_The_Emperor=e2=80=99s_New_Clot?= =?UTF-8?Q?hes_[John_Sowa]=29?= Date: Sun, 5 Jan 2025 21:46:33 +0100 Message-ID: References: MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit Injection-Date: Sun, 5 Jan 2025 20:46:29 -0000 (UTC) Injection-Info: solani.org; logging-data="2377896"; mail-complaints-to="abuse@news.solani.org" User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:91.0) Gecko/20100101 Firefox/91.0 SeaMonkey/2.53.19 Cancel-Lock: sha1:ABL2B3K4wrgcIAyJ9J4WBuF/5Wk= In-Reply-To: X-User-ID: eJwFwYEBwCAIA7CXFGid5xSE/09YAudmnSAYGIx/LlOZLnb4km4yz3Dak/XWm+4otbZwYoQpSyHSSkmrH3YeFx4= Bytes: 2786 Lines: 49 Notice John Sowa calls LLM the “store” of GPT. This could be a misconception that matches what Permion did for their cognitive memory. But matters are a little bit more complicated to say the least, especially since OpenAI insists that GPT itself is also an LLM. What might highlight the situation is Fig 6 of this paper, postulating two Mixture of Experts (MoE), one on attention mechanism and one on feed-forward: A Survey on Mixture of Experts [2407.06204] A Survey on Mixture of Experts https://arxiv.org/abs/2407.06204 Disclaimer: Pitty Marvin Minksy didn’t describe these things already in his society of mind! Would make it easier to understand it now… Mild Shock schrieb: > Douglas Lenat died two years ago in > August 31, 2023. I don’t know whether > CYC and Cycorp will make a dent in > the future. CYC adressed the common > > knowledge bottleneck, and so do LLM. I > am using CYC mainly as a historical reference. > The “common knowledge bottleneck” in AI is > a challenge that plagued early AI systems. > This bottleneck stems from the difficulty > > of encoding vast amounts of everyday, > implicit human knowledge things we take for > granted but computers historically struggled > to understand. Currently LLM by design focus > more on shallow > > knowledge, whereas systems such as CYC might > exhibit more deep knowlege in certain domains, > making them possibly more suitable when the > stakeholders expect more reliable > analytic capabilities. > > The problem is not explainability, > the problem is intelligence. > >