Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Evaluating Complex Sparse Representation of Hypervectors for Unsupervised Machine Learning
Centre for Data Analytics and Cognition at La Trobe University, Melbourne, Australia.
Luleå tekniska universitet, Institutionen för system- och rymdteknik, Datavetenskap.ORCID-id: 0000-0003-0069-640x
Centre for Data Analytics and Cognition at La Trobe University, Melbourne, Australia.
Centre for Data Analytics and Cognition at La Trobe University, Melbourne, Australia.
Visa övriga samt affilieringar
2022 (Engelska)Ingår i: 2022 International Joint Conference on Neural Networks (IJCNN): 2022 Conference Proceedings, IEEE, 2022Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

The increasing use of Vector Symbolic Architectures (VSA) in machine learning has contributed towards en-ergy efficient computation, short training cycles and improved performance. A further advancement of VSA is to leverage sparse representations, where the VSA-encoded hypervectors are sparsified to represent receptive field properties when encoding sensory inputs. The hyperseed algorithm is an unsupervised machine learning algorithm based on VSA for fast learning a topology preserving feature map of unlabelled data. In this paper, we implement two methods of sparse block-codes on the hyperseed algorithm, they are selecting the maximum element of each block and selecting a random element of each block as the nonzero element. Finally, the sparsified hyperseed algorithm is empirically evaluated for performance using three distinct bench-mark datasets, Iris classification, classification and visualisation of synthetic datasets from the Fundamental Clustering Problems Suite and language classification using n-gram statistics.

Ort, förlag, år, upplaga, sidor
IEEE, 2022.
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
Kommunikations- och beräkningssystem
Identifikatorer
URN: urn:nbn:se:ltu:diva-94787DOI: 10.1109/IJCNN55064.2022.9892981ISI: 000867070908091Scopus ID: 2-s2.0-85140777444OAI: oai:DiVA.org:ltu-94787DiVA, id: diva2:1717489
Konferens
IEEE World Congress on Computational Intelligence (WCCI 2022), International Joint Conference on Neural Networks (IJCNN 2022), Padua, Italy, July 18-23, 2022
Anmärkning

ISBN för värdpublikation: 978-1-7281-8671-9

Tillgänglig från: 2022-12-08 Skapad: 2022-12-08 Senast uppdaterad: 2025-10-21Bibliografiskt granskad

Open Access i DiVA

Fulltext saknas i DiVA

Övriga länkar

Förlagets fulltextScopus

Person

Osipov, Evgeny

Sök vidare i DiVA

Av författaren/redaktören
Osipov, Evgeny
Av organisationen
Datavetenskap
Datavetenskap (datalogi)

Sök vidare utanför DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetricpoäng

doi
urn-nbn
Totalt: 180 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf