3260 papers • 126 benchmarks • 313 datasets
Entity Embeddings is a technique for applying deep learning to tabular data. It involves representing the categorical data of an information systems entity with multiple dimensions.
(Image credit: Papersgraph)
These leaderboards are used to track progress in entity-embeddings-10
No benchmarks available.
Use these libraries to find entity-embeddings-10 models and implementations
No subtasks available.
It is demonstrated in this paper that entity embedding helps the neural network to generalize better when the data is sparse and statistics is unknown, and is especially useful for datasets with lots of high cardinality features, where other methods tend to overfit.
This paper introduces a conceptually simple, scalable, and highly effective BERT-based entity linking model, along with an extensive evaluation of its accuracy-speed trade-off. We present a two-stage zero-shot linking algorithm, where each entity is defined only by a short textual description. The first stage does retrieval in a dense space defined by a bi-encoder that independently embeds the mention context and the entity descriptions. Each candidate is then re-ranked with a cross-encoder, that concatenates the mention and entity text. Experiments demonstrate that this approach is state of the art on recent zero-shot benchmarks (6 point absolute gains) and also on more established non-zero-shot evaluations (e.g. TACKBP-2010), despite its relative simplicity (e.g. no explicit entity embeddings or manually engineered mention tables). We also show that bi-encoder linking is very fast with nearest neighbor search (e.g. linking with 5.9 million candidates in 2 milliseconds), and that much of the accuracy gain from the more expensive cross-encoder can be transferred to the bi-encoder via knowledge distillation. Our code and models are available at https://github.com/facebookresearch/BLINK.
This work proposes a deep learning based content-collaborative methodology for personalized size and fit recommendation that can ingest arbitrary customer and article data and can model multiple individuals or intents behind a single account.
Know-Evolve is presented, a novel deep evolutionary knowledge network that learns non-linearly evolving entity representations over time that effectively predicts occurrence or recurrence time of a fact which is novel compared to prior reasoning approaches in multi-relational setting.
This work proposes HyperTeNet - a self-attention hypergraph and Transformer-based neural network architecture for the personalized list continuation task that significantly outperforms the other state-of-the-art models on real-world datasets.
It is suggested that existing GCNs are unnecessary for KGC, and novel GCN-based KGC models should count on more ablation studies to validate their effectiveness.
ClusterEA is presented, a general framework that is capable of scaling up EA models and enhancing their results by leveraging normalization methods on mini-batches with a high entity equivalent rate and contains three components to align entities between large-scale KGs, including stochastic training, ClusterSampler, and SparseFusion.
The experimental results proves that StarGraph is efficient in parameters, and the improvement made on ogbl-wikikg2 demonstrates its great effectiveness of representation learning on large-scale knowledge graphs.
This work devise a differentiable vector quantization framework for automatic category tree generation, namely CAGE, which enables the simultaneous learning and refinement of categorical code representations and entity embeddings in an end-to-end manner, starting from the randomly initialized states.
A neural model is designed and trained with a novel method for sampling informative negative examples that significantly outperforms existing state-of-the-art methods on WikilinksNED while achieving comparable performance on a smaller newswire dataset.
Adding a benchmark result helps the community track progress.