18 Matching Annotations
  1. May 2025
    1. The answer most technocrats are leaning towards is vector search technology and Retrieval-Augmented Generation (RAG) models that improve AI experiences. These intelligent search systems are fundamentally changing how users discover information, interact with applications, and receive personalized experiences across industries.

      Explore how embedding intelligence transforms Vector Search and RAG (Retrieval-Augmented Generation) models. Learn the key benefits, use cases, and implementation strategies for smarter AI-driven search systems.

  2. Feb 2025
    1. Las incrustaciones de palabras, o word embeddings, son un método que consiste en derivar, por medio de algoritmos diversos, representaciones numéricas vectoriales de los términos presentes en un corpus, de acuerdo con sus relaciones contextuales7Michael Gavin et al., «Spaces of Meaning: Conceptual HIstory, Vector Semantics, and Close Reading», Debates in the Digital Humanities 2019, ed. Matthew K. Gold y Lauren F. Klein (Minneapolis London: University of Minnesota Press, 2019), 243-67.. Este método sigue los principios de la semántica distribucional8
  3. Jul 2021
    1. Recommendations DON'T use shifted PPMI with SVD. DON'T use SVD "correctly", i.e. without eigenvector weighting (performance drops 15 points compared to with eigenvalue weighting with (p = 0.5)). DO use PPMI and SVD with short contexts (window size of (2)). DO use many negative samples with SGNS. DO always use context distribution smoothing (raise unigram distribution to the power of (lpha = 0.75)) for all methods. DO use SGNS as a baseline (robust, fast and cheap to train). DO try adding context vectors in SGNS and GloVe.
  4. Jun 2020
  5. Dec 2019
    1. The quality of word representations is generally gauged by its ability to encode syntactical information and handle polysemic behavior (or word senses). These properties result in improved semantic word representations. Recent approaches in this area encode such information into its embeddings by leveraging the context. These methods provide deeper networks that calculate word representations as a function of its context.
      • Syntactical information
      • Polysemic behavior (word senses)
      • Semantic word representations

      Entendo que lidar com word senses significa dizer que a representação das palavras consegue medidas similares para palavras similares.

      O que seria informação sintática? E sua relação com representações semânticas da palavra?

    2. Traditional word embedding algorithms assign a distinct vector to each word. This makes them unable to account for polysemy. In a recent work, Upadhyay et al. (2017) provided an innovative way to address this deficit. The authors leveraged multilingual parallel data to learn multi-sense word embeddings.
      • multilingual parallel data
      • multi-sense word embeddings
    3. This is very important as training embeddings from scratch requires large amount of time and resource. Mikolov et al. (2013) tried to address this issue by proposing negative sampling which is nothing but frequency-based sampling of negative terms while training the word2vec model.

      Amostragem negativa... termos negativos?

    4. A general caveat for word embeddings is that they are highly dependent on the applications in which it is used. Labutov and Lipson (2013) proposed task specific embeddings which retrain the word embeddings to align them in the current task space.

      Acredito que aplicação aqui se relaciona com contexto, logo word embeddings são dependentes de contexto. Isso é bem óbvio, a princípio. Seria isso o que o autor quis dizer?

      Retreinar as incorporações para alinhar à tarefa corrente. Alinhar seria nada mais do que adequar as incorporações prévias no novo contexto, é isso?

    5. One solution to this problem, as explored by Mikolov et al. (2013), is to identify such phrases based on word co-occurrence and train embeddings for them separately. More recent methods have explored directly learning n-gram embeddings from unlabeled data (Johnson and Zhang, 2015).

      Co-ocorrência de palavras eu consigo entender, mas treinar as embeddings separadamente não. Seria supor a co-ocorrência das palavras como unidade na incorporação, em vez da palavra apenas?

  6. Jun 2017
  7. Apr 2017
  8. Jun 2016