Nested Learning: Part II

#ai #llm #deeplearning #ml

Here is the video lecture, subscribe for updates,

https://www.youtube.com/watch?v=qDwAm1f0JBY

Some concerns, it catches the forgetting behavior of LLMs, the context once gone can be looked into nested memories and hence can be revised if nested parameters and memories are referenced. The pre training can’t be edited but the computation of temporary context can be referred if intact in these layers.

The terminologies and formulas we use are here, are taken from Reference [1].

Associated Memories can be defined as follows,

Reference [1]

Reference [1]

The same thing, but in key-value-based terminologies of associated memories:

Reference [1]

Reference [1]

Reference [1]

Reference [1]

Reference [1]

Reference [1]

The experimental results are good and have been explained in [1].

References

[1] Behrouz, A., Razaviyayn, M., Zhong, P., & Mirrokni, V. (2025). Nested learning: The illusion of deep learning architectures. In The Thirty-ninth Annual Conference on Neural Information Processing Systems.

[2] https://youtu.be/qDwAm1f0JBY

Published by Nidhika

Hi, Apart from profession, I have inherent interest in writing especially about Global Issues of Concern, fiction blogs, poems, stories, doing painting, cooking, photography, music to mention a few! And most important on this website you can find my suggestions to latest problems, views and ideas, my poems, stories, novels, some comments, proposals, blogs, personal experiences and occasionally very short glimpses of my research work as well.

Leave a comment