Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Recursive Binding for Similarity-Preserving Hypervector Representations of Sequences
Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Computer Science. International Research and Training Center for Information Technologies and Systems, Kiev, Ukraine.ORCID iD: 0000-0002-3414-5334
University of California at Berkeley, Redwood Center for Theoretical Neuroscience, Berkeley, USA; Research Institutes of Sweden, Intelligent Systems Lab, Kista, Sweden.
2022 (English)In: 2022 International Joint Conference on Neural Networks (IJCNN): 2022 Conference Proceedings, IEEE, 2022Conference paper, Published paper (Refereed)
Abstract [en]

Hyperdimensional computing (HDC), also known as vector symbolic architectures (VSA), is a computing framework used within artificial intelligence and cognitive computing that operates with distributed vector representations of large fixed dimensionality. A critical step in designing the HDC/VSA solutions is to obtain such representations from the input data. Here, we focus on a wide-spread data type of sequences and propose their transformation to distributed representations that both preserve the similarity of identical sequence elements at nearby positions and are equivariant with respect to the sequence shift. These properties are enabled by forming representations of sequence positions using recursive binding as well as superposition operations. The proposed transformation was experimentally investigated with symbolic strings used for modeling human perception of word similarity. The obtained results are on a par with more sophisticated approaches from the literature. The proposed transformation was designed for the HDC/VSA model known as Fourier Holographic Reduced Representations. However, it can be adapted to some other HDC/VSA models.

Place, publisher, year, edition, pages
IEEE, 2022.
Keywords [en]
data structures, distributed representation, hyperdimensional computing, hypervector, recursive binding, sequence representation, shift equivariance, similarity preserving transformation, vector symbolic architectures
National Category
Computer Sciences Information Systems
Research subject
Dependable Communication and Computation Systems
Identifiers
URN: urn:nbn:se:ltu:diva-94150DOI: 10.1109/IJCNN55064.2022.9892462ISI: 000867070904096Scopus ID: 2-s2.0-85137519384OAI: oai:DiVA.org:ltu-94150DiVA, id: diva2:1711876
Conference
IEEE World Congress on Computational Intelligence (WCCI 2022), International Joint Conference on Neural Networks (IJCNN 2022), Padua, Italy, July 18-23, 2022
Funder
EU, Horizon 2020, 839179Swedish Foundation for Strategic Research, UKR22-0024
Note

Funder: AFOSR (FA9550-19-1-0241), Intel’s THWAI program, National Academy of Sciences of Ukraine (0121U000016), Ministry of Education and Science of Ukraine (0121U000228, 0122U000818);

ISBN för värdpublikation: 978-1-7281-8671-9

Available from: 2022-11-18 Created: 2022-11-18 Last updated: 2023-05-08Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Rachkovskij, Dmitri A.

Search in DiVA

By author/editor
Rachkovskij, Dmitri A.
By organisation
Computer Science
Computer SciencesInformation Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 100 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf