Deutsch   English   Français   Italiano  
<vk0pph$12ke2$2@solani.org>

View for Bookmarking (what is this?)
Look up another Usenet article

Path: ...!news.roellig-ltd.de!open-news-network.org!weretis.net!feeder8.news.weretis.net!reader5.news.weretis.net!news.solani.org!.POSTED!not-for-mail
From: Mild Shock <janburse@fastmail.fm>
Newsgroups: comp.lang.prolog
Subject: What does "o1" mean in recent Models (Re: Ilya Sutskever: The Next
 Oppenheimer)
Date: Thu, 19 Dec 2024 10:42:10 +0100
Message-ID: <vk0pph$12ke2$2@solani.org>
References: <vjumvu$1es1e$2@solani.org>
MIME-Version: 1.0
Content-Type: text/plain; charset=UTF-8; format=flowed
Content-Transfer-Encoding: 7bit
Injection-Date: Thu, 19 Dec 2024 09:42:09 -0000 (UTC)
Injection-Info: solani.org;
	logging-data="1135042"; mail-complaints-to="abuse@news.solani.org"
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:91.0) Gecko/20100101
 Firefox/91.0 SeaMonkey/2.53.19
Cancel-Lock: sha1:Ea+58eotYQA8vCsN7RJjjd+aATI=
X-User-ID: eJwVwoENwDAIA7CX1kJCOaeE8f8J0yzDuKhwgo7BvF22H9OfBULVrrf9duUErU9o2DsVOWtjEXVkGRbHdD9sURXN
In-Reply-To: <vjumvu$1es1e$2@solani.org>
Bytes: 2222
Lines: 41

Hi,

Could it be that "o1" likely refers to "Optimizer 1".
And what could this include?

- Compressing weights or activations into
fewer bits can significantly reduce computation,
especially in hardware, mimicking O(1)-like
efficiency for certain operations.

- Removing redundant connections in the neural
network leads to fewer computations. Sparse matrix
operations can optimize dense workloads, making specific
inference tasks faster.

- Large models are distilled into smaller ones
with similar capabilities, reducing computational
costs during inference. If the optimized paths are
cleverly structured, their complexity might be
closer to O(1) for lookup-style tasks.

So maybe Ilya Sutskever wants to tell us, in his
recent talk when he refers to the 700g brain line:
Look we did the same as biological evolution,

we found a way to construct more compact brains.

Bye

Mild Shock schrieb:
> Hi,
> 
> I liked some videos on YouTube:
> 
> Ilya Sutskever: The Next Oppenheimer
> https://www.youtube.com/watch?v=jryDWOKikys
> 
> Ilya Sutskever: Sequence to Sequence Learning
> https://www.youtube.com/watch?v=WQQdd6qGxNs
> 
> Bye