Probabilistic Models

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 77178 Experts worldwide ranked by ideXlab platform

Olli Simula - One of the best experts on this subject based on the ideXlab platform.

  • EUSIPCO - A learning vector quantization algorithm for Probabilistic Models
    2000
    Co-Authors: Jaakko Hollmén, Volker Tresp, Olli Simula
    Abstract:

    In classification problems, it is preferred to attack the discrimination problem directly rather than indirectly by first estimating the class densities and by then estimating the discrimination function from the generative Models through Bayes's rule. Sometimes, however, it is convenient to express the Models as Probabilistic Models, since they are generative in nature and can handle the representation of high-dimensional data like time-series. In this paper, we derive a discriminative training procedure based on Learning Vector Quantization (LVQ) where the codebook is expressed in terms of Probabilistic Models. The likelihood-based distance measure is justified using the Kullback-Leibler distance. In updating the winner unit, a gradient learning step is taken with regard to the parameters of the Probabilistic model. The method essentially departs from a prototypical representation and incorporates learning in the parameter space of generative Models. As an illustration, we present experiments in the fraud detection domain, where Models of calling behavior are used to classify mobile phone subscribers to normal and fraudulent users. This is an extension of our earlier work in clustering Probabilistic Models with the Self-Organizing Map (SOM) algorithm to the classification domain.

  • A learning vector quantization algorithm for Probabilistic Models
    2000 10th European Signal Processing Conference, 2000
    Co-Authors: Jaakko Hollmén, Volker Tresp, Olli Simula
    Abstract:

    In classification problems, it is preferred to attack the discrimination problem directly rather than indirectly by first estimating the class densities and by then estimating the discrimination function from the generative Models through Bayes's rule. Sometimes, however, it is convenient to express the Models as Probabilistic Models, since they are generative in nature and can handle the representation of high-dimensional data like time-series. In this paper, we derive a discriminative training procedure based on Learning Vector Quantization (LVQ) where the codebook is expressed in terms of Probabilistic Models. The likelihood-based distance measure is justified using the Kullback-Leibler distance. In updating the winner unit, a gradient learning step is taken with regard to the parameters of the Probabilistic model. The method essentially departs from a prototypical representation and incorporates learning in the parameter space of generative Models. As an illustration, we present experiments in the fraud detection domain, where Models of calling behavior are used to classify mobile phone subscribers to normal and fraudulent users. This is an extension of our earlier work in clustering Probabilistic Models with the Self-Organizing Map (SOM) algorithm to the classification domain.

Steve Lawrence - One of the best experts on this subject based on the ideXlab platform.

  • Probabilistic Models for unified collaborative and content based recommendation in sparse data environments
    arXiv: Information Retrieval, 2013
    Co-Authors: Alexandrin Popescul, Lyle H Ungar, David M Pennock, Steve Lawrence
    Abstract:

    Recommender systems leverage product and community information to target products to consumers. Researchers have developed collaborative recommenders, content-based recommenders, and (largely ad-hoc) hybrid systems. We propose a unified Probabilistic framework for merging collaborative and content-based recommendations. We extend Hofmann's [1999] aspect model to incorporate three-way co-occurrence data among users, items, and item content. The relative influence of collaboration data versus content data is not imposed as an exogenous parameter, but rather emerges naturally from the given data sources. Global Probabilistic Models coupled with standard Expectation Maximization (EM) learning algorithms tend to drastically overfit in sparse-data situations, as is typical in recommendation applications. We show that secondary content information can often be used to overcome sparsity. Experiments on data from the ResearchIndex library of Computer Science publications show that appropriate mixture Models incorporating secondary data produce significantly better quality recommenders than k-nearest neighbors (k-NN). Global Probabilistic Models also allow more general inferences than local methods like k-NN.

  • Probabilistic Models for unified collaborative and content based recommendation in sparse data environments
    Uncertainty in Artificial Intelligence, 2001
    Co-Authors: Alexandrin Popescul, Lyle H Ungar, David M Pennock, Steve Lawrence
    Abstract:

    Recommender systems leverage product and community information to target products to consumers. Researchers have developed collaborative recommenders, content-based recommenders, and a few hybrid systems. We propose a unified Probabilistic framework for merging collaborative and content-based recommendations. We extend Hofmarm's (1999) aspect model to incorporate three-way co-occurrence data among users, items, and item content. The relative influence of collaboration data versus content data is not imposed as an exogenous parameter, but rather emerges naturally from the given data sources. However, global Probabilistic Models coupled with standard EM learning algorithms tend to drastically overfit in the sparsedata situations typical of recommendation applications. We show that secondary content information can often be used to overcome sparsity. Experiments on data from the Researchlndex library of Computer Science publications show that appropriate mixture Models incorporating secondary data produce significantly better quality recommenders than k-nearest neighbors (k-NN). Global Probabilistic Models also allow more general inferences than local methods like k-NN.

Fernando Pereira - One of the best experts on this subject based on the ideXlab platform.

  • conditional random fields Probabilistic Models for segmenting and labeling sequence data
    International Conference on Machine Learning, 2001
    Co-Authors: John Lafferty, Andrew Mccallum, Fernando Pereira
    Abstract:

    We present conditional random fields , a framework for building Probabilistic Models to segment and label sequence data. Conditional random fields offer several advantages over hidden Markov Models and stochastic grammars for such tasks, including the ability to relax strong independence assumptions made in those Models. Conditional random fields also avoid a fundamental limitation of maximum entropy Markov Models (MEMMs) and other discriminative Markov Models based on directed graphical Models, which can be biased towards states with few successor states. We present iterative parameter estimation algorithms for conditional random fields and compare the performance of the resulting Models to HMMs and MEMMs on synthetic and natural-language data.

Jaakko Hollmén - One of the best experts on this subject based on the ideXlab platform.

  • EUSIPCO - A learning vector quantization algorithm for Probabilistic Models
    2000
    Co-Authors: Jaakko Hollmén, Volker Tresp, Olli Simula
    Abstract:

    In classification problems, it is preferred to attack the discrimination problem directly rather than indirectly by first estimating the class densities and by then estimating the discrimination function from the generative Models through Bayes's rule. Sometimes, however, it is convenient to express the Models as Probabilistic Models, since they are generative in nature and can handle the representation of high-dimensional data like time-series. In this paper, we derive a discriminative training procedure based on Learning Vector Quantization (LVQ) where the codebook is expressed in terms of Probabilistic Models. The likelihood-based distance measure is justified using the Kullback-Leibler distance. In updating the winner unit, a gradient learning step is taken with regard to the parameters of the Probabilistic model. The method essentially departs from a prototypical representation and incorporates learning in the parameter space of generative Models. As an illustration, we present experiments in the fraud detection domain, where Models of calling behavior are used to classify mobile phone subscribers to normal and fraudulent users. This is an extension of our earlier work in clustering Probabilistic Models with the Self-Organizing Map (SOM) algorithm to the classification domain.

  • A learning vector quantization algorithm for Probabilistic Models
    2000 10th European Signal Processing Conference, 2000
    Co-Authors: Jaakko Hollmén, Volker Tresp, Olli Simula
    Abstract:

    In classification problems, it is preferred to attack the discrimination problem directly rather than indirectly by first estimating the class densities and by then estimating the discrimination function from the generative Models through Bayes's rule. Sometimes, however, it is convenient to express the Models as Probabilistic Models, since they are generative in nature and can handle the representation of high-dimensional data like time-series. In this paper, we derive a discriminative training procedure based on Learning Vector Quantization (LVQ) where the codebook is expressed in terms of Probabilistic Models. The likelihood-based distance measure is justified using the Kullback-Leibler distance. In updating the winner unit, a gradient learning step is taken with regard to the parameters of the Probabilistic model. The method essentially departs from a prototypical representation and incorporates learning in the parameter space of generative Models. As an illustration, we present experiments in the fraud detection domain, where Models of calling behavior are used to classify mobile phone subscribers to normal and fraudulent users. This is an extension of our earlier work in clustering Probabilistic Models with the Self-Organizing Map (SOM) algorithm to the classification domain.

Alexandrin Popescul - One of the best experts on this subject based on the ideXlab platform.

  • Probabilistic Models for unified collaborative and content based recommendation in sparse data environments
    arXiv: Information Retrieval, 2013
    Co-Authors: Alexandrin Popescul, Lyle H Ungar, David M Pennock, Steve Lawrence
    Abstract:

    Recommender systems leverage product and community information to target products to consumers. Researchers have developed collaborative recommenders, content-based recommenders, and (largely ad-hoc) hybrid systems. We propose a unified Probabilistic framework for merging collaborative and content-based recommendations. We extend Hofmann's [1999] aspect model to incorporate three-way co-occurrence data among users, items, and item content. The relative influence of collaboration data versus content data is not imposed as an exogenous parameter, but rather emerges naturally from the given data sources. Global Probabilistic Models coupled with standard Expectation Maximization (EM) learning algorithms tend to drastically overfit in sparse-data situations, as is typical in recommendation applications. We show that secondary content information can often be used to overcome sparsity. Experiments on data from the ResearchIndex library of Computer Science publications show that appropriate mixture Models incorporating secondary data produce significantly better quality recommenders than k-nearest neighbors (k-NN). Global Probabilistic Models also allow more general inferences than local methods like k-NN.

  • Probabilistic Models for unified collaborative and content based recommendation in sparse data environments
    Uncertainty in Artificial Intelligence, 2001
    Co-Authors: Alexandrin Popescul, Lyle H Ungar, David M Pennock, Steve Lawrence
    Abstract:

    Recommender systems leverage product and community information to target products to consumers. Researchers have developed collaborative recommenders, content-based recommenders, and a few hybrid systems. We propose a unified Probabilistic framework for merging collaborative and content-based recommendations. We extend Hofmarm's (1999) aspect model to incorporate three-way co-occurrence data among users, items, and item content. The relative influence of collaboration data versus content data is not imposed as an exogenous parameter, but rather emerges naturally from the given data sources. However, global Probabilistic Models coupled with standard EM learning algorithms tend to drastically overfit in the sparsedata situations typical of recommendation applications. We show that secondary content information can often be used to overcome sparsity. Experiments on data from the Researchlndex library of Computer Science publications show that appropriate mixture Models incorporating secondary data produce significantly better quality recommenders than k-nearest neighbors (k-NN). Global Probabilistic Models also allow more general inferences than local methods like k-NN.