Item request has been placed! ×
Item request cannot be made. ×
loading  Processing Request

Structure and Representation in Neural Networks

Item request has been placed! ×
Item request cannot be made. ×
loading   Processing Request
  • معلومة اضافية
    • Contributors:
      Russell, Stuart
    • بيانات النشر:
      eScholarship, University of California
    • الموضوع:
      2024
    • Collection:
      University of California: eScholarship
    • نبذة مختصرة :
      Since neural networks have become dominant within the field of artificial intelligence, a sub-field of research has emerged attempting to understand their inner workings. One standard method within this sub-field has been to primarily understand neural networks as representing human-comprehensible features. Another possibility that has been less explored is to understand them as multi-step computer programs. A seeming prerequisite for this is some form of modularity: for different parts of the network to operate independently enough to be understood in isolation, and to implement distinct interpretable sub-routines.To find modular structure inside neural networks, we initially use the tools of graphical clustering. A network is clusterable in this sense if it can be divided into groups of neurons with strong internal connectivity but weak external connectivity. We find that a trained neural network is typically more clusterable than randomly initialized networks, and often clusterable relative to random networks with the same distribution of weights as the trained network. We investigate factors that promote clusterability, and also develop novel methods targeted at that end.For modularity to be valuable for understanding neural networks, it needs to have some sort of functional relevance.The type of functional relevance we target is local specialization of functionality. A neural network is locally specialized to the extent that parts of its computational graph can be abstractly represented as performing some comprehensible sub-task relevant to the overall task. We propose two proxies for local specialization: importance, which reflects how valuable sets of neurons are to network performance; and coherence, which reflects how consistently their neurons associate with features of the inputs. We then operationalize these proxies using techniques conventionally used to interpret individual neurons, applying them instead to groups of neurons produced by graph clustering algorithms. Our results show that clustering ...
    • File Description:
      application/pdf
    • Relation:
      qt3qb142b4; https://escholarship.org/uc/item/3qb142b4; https://escholarship.org/content/qt3qb142b4/qt3qb142b4.pdf
    • Rights:
      public
    • الرقم المعرف:
      edsbas.940F751D