Code for the Information Retrieval course of the MSc in Artificial Intelligence at the University of Amsterdam.
One of the key questions in IR is whether offline evaluation and online evaluation outcomes agree with each other. We study the degree of agreement between offline evaluation measures and interleaving outcomes, by the means of simulations using several click models.
In a typical IR task we are interested in finding a (usually ranked) list of results that satisfy the information need of a user expressed by means of a query. Many difficulties arise as the satisfaction and information need of the user cannot be directly observed, and thus the relevance of a particular document given the query is unobserved. Moreover, the query is merely a linguistic representation of the actual information need of the user and the gap between them can not be measured either. We explore three families of models that are used to measure and rank the relevance of a set of documents given a query: lexical models, semantic models, and machine learning-based re-ranking algorithms that build on top of the former models.
We implement and compare the performance of several approaches to the Learning to Rank (LTR) problem. We received an implementation of a pointwise method based on squared-loss minimization and implemented RankNet (pairwise) and LambdaNet (listwise) algorithms.
Refer to each lab and run the iPython notebook as follows.
jupyter notebook $notebook_name$.ipynb
- Matplotlib
- NumPy
- pandas
- SciPy
- Pyndri
- Gensim
- Lasagne
- Theano
Copyright © 2017 Dana Kianfar and Jose Gallego.
This project is distributed under the MIT license. This was developed as part of the Information Retrieval course taught by Evangelos Kanoulas at the University of Amsterdam. Please review the UvA regulations governing Fraud and Plagiarism in case you are a student at the UvA.