On Design Choices in Similarity-Preserving Sparse Randomized Embeddings
2024 (English)In: 2024 International Joint Conference on Neural Networks (IJCNN), IEEE, 2024Conference paper, Published paper (Refereed)
Abstract [en]
Expand & Sparsify is a principle that is observed in anatomically similar neural circuits found in the mushroom body (insects) and the cerebellum (mammals). Sensory data are projected randomly to much higher-dimensionality (expand part) where only few the most strongly excited neurons are activated (sparsify part). This principle has been leveraged to design a FlyHash algorithm that forms similarity-preserving sparse embeddings, which have been found useful for such tasks as novelty detection, pattern recognition, and similarity search. Despite its simplicity, FlyHash has a number of design choices to be set such as preprocessing of the input data, choice of sparsifying activation function, and formation of the random projection matrix. In this paper, we explore the effect of these choices on the performance of similarity search with FlyHash embeddings. We find that the right combination of design choices can lead to drastic difference in the search performance.
Place, publisher, year, edition, pages
IEEE, 2024.
Keywords [en]
random projection, Winner-Take-All, sparse representations, hyperdimensional computing, expand & sparsify
National Category
Computer Systems
Research subject
Dependable Communication and Computation Systems
Identifiers
URN: urn:nbn:se:ltu:diva-110298DOI: 10.1109/IJCNN60899.2024.10651277ISI: 001392668204066Scopus ID: 2-s2.0-85205027306OAI: oai:DiVA.org:ltu-110298DiVA, id: diva2:1904774
Conference
13th IEEE World Congress on Computational Intelligence (WCCI 2024), Yokohama, Japan, June 30 - July 5, 2024
Note
Funder: Horizon 2020 (839179); Swedish Foundation for Strategic Research (UKR22-0024 & UKR24-0014); Swedish Research Council Scholars at Risk Sweden (2022/1963);
ISBN for host publication: 978-8-3503-5931-2;
2024-10-102024-10-102025-10-21Bibliographically approved