Purpose. An introduction to information retrieval including indexing, retrieval, classifying, and clustering text and multimedia documents.
Book. Introduction to Information Retrieval by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze
Evaluation. Homework/lab projects (1/2) + Summaries (1/4) + Quizzes (1/4)
- Lectures cover the material in the reading materials by placing it in
context, giving examples, and engaging in Q&As.
Instructor: Prof. Cristina Lopes, DBH 5076, lopes
at ics dot uci
Office hours: Mondays 1pm-2pm; Wednesdays 11am-12pm. DBH 5076
Readers: Wen Shen and Oliver Wang
Lectures: Mon & Wed 3:30-4:50pm, SH 134
There will be 4 projects, the first one being trivial and the last one having several milestones with their own deadlines. Projects are due by midnight on the due date. Late projects will be accepted with penalties -- see course policies.
See instructions in each project description.
|0||Course policies: read, sign, scan (or take picture) and upload to EEE||1/11||1%|
|3||Search Engine||2/18, 2/28, 3/14||60%|
There will be 4 quizzes throughout the course. Quizzes are on Mondays during the lecture. They cover material that has been taught the previous weeks since the last quizz. The quiz with the worst score will be discarded. No quiz make-ups.
Summaries are due Saturdays. Summaries submitted up to one week late will have a penalty of 25%. No summaries will be accepted past 1 week of their due date.
Each article should be summarized in no more than one page, with the following structure: (a) objective summary of the article (do not inject your views here, be objective); (b) short personal commentary about the article (your views here).
Submit one pdf file per week with all the summaries for that week on that file.
Please name your paper summary files like this: LastName_WeekNumber.pdf starting with WeekNumber=1 for the first week. Files that don't follow this convention may be missed by the instructors.
Include your full name and student ID in the summary itself.
Turn in summaries in EEE Dropbox.
|1||1/5||Web Search Basics||Textbook Chapter
19: Web Search Basics (no need to summarize)
1. Wikipedia entry on Vannevar Bush
2. "As We May Think" The Atlantic Monthly, July, 1945. (reprinted in ACM CHI Interactions, March 1996) (*)
Search Engine Optimization
3. "Stuff I've seen: A system for personal information retrieval and re-use " by S. Dumais, E. Cutell, J. Cadiz, G. Jancke, R. Sarin, and D. Robbins, SIGIR, 2003
Commentary: "This paper addresses an increasingly important problem - how to search and manage personal collections of electronic information. ... it addresses an important user-centered problem. ...this paper presents a practical user interface to make the system useful. ..., the paper includes large scale, user-oriented testing that demonstrates the efficacy of the system. ..., the evaluation uses both quantitative and qualitative data to make its case. I think this paper is destined to be a classic because it may eventually define how people manage their files for a decade. Moreover, it is well-written and can serve as a good model for developers doing system design and evaluation, and for students learning about IR systems and evaluation."
4. "Simple, Proven Approaches to Text Retrieval" by Robertson and Jones (*)
Commentary: "This paper provides a brief but well informed and technically accurate overview of the state of the art in text retrieval, at least up to 1997. It introduces the ideas of terms and matching, term weighting strategies, relevance weighting, a little on data structures and the evidence for their effectiveness. In my view it does an exemplary job of introducing the terminology of IR and the main issues in text retrieval for a numerate and technically well informed audience. It also has a very well chosen list of references."
|3||1/19*||Web crawling||Textbook Chapter
20 : Web Crawling and Indices (no need to summarize)
5. "The Web As a Graph" by R. Kumar, P Raghavan, S. Rajagopalan, D. Sivakumar, A. Tomkins, E. Upfal, PODS 2000 (*)
Abstract: "The pages and hyperlinks of the World-Wide Web may be viewed as nodes and edges in a directed graph. This graph has about a billion nodes today, several billion links, and appears to grow exponentially with time. There are many reasons -- mathematical, sociological, and commercial -- for studying the evolution of this graph. We first review a set of algorithms that operate on the Web graph, addressing problems from Web search, automatic community discovery, and classification. We then recall a number of measurements and properties of the Web graph. Noting that traditional random graph models do not explain these observations, we propose a new family of random graph models."
6. "How Google Code Search Worked
" by Russ Cox (January 2012) (*)
|5||2/2||Index Construction and Scoring||
Prof. Chen Li will share his experiences of building a commercial enterprise search engine at his startup SRCH2.Textbook Chapter 4 : Index Construction
Commentary: Three Google researchers summarize the benefits of data-driven problem-solving in an essay that borrows the title from another famous paper that proposes the opposite.
|6||2/9||Querying, Scoring, Term Weighting and the Vector Space model||Textbook Chapter
1 : Boolean Retrieval
Textbook Chapter 6 : Scoring, term weighting & the vector space model
8.A vector space model for automatic indexing by Salton, Wong, Yang (*)
|7||2/16*||Search Engine Evaluation
Vector Space Model
9. "Map Reduce: Simplified Data Processing on Large Clusters" by Jeffrey Dean and Sanjay Ghemawat (*)
Commentary: the paper that revolutionized modern data processing, made "cloud computing" trendy, and a great example of how programming language concepts can be applied to the design of real systems.
Textbook Chapter 21 : Link Analysis
10. "The Anatomy of a Large-Scale Hypertextual Web Search Engine" by S. Brin and L. Page (this link is to the long version, the short version was publishied in WWW1998)(*)
Commentary: "This paper (and the work it reports) has had more impact on everyday life than any other in the IR area. A major contribution of the paper is the recognition that some relevant search results are greatly more valued by searchers than others. By reflecting this in their evaluation procedures, Brin and Page were able to see the true value of web-specific methods like anchor text. The paper presents a highly efficient, scalable implementation of a ranking method which now delivers very high quality results to a billion people over billions of pages at about 6,000 queries per second. It also hints at the technology which Google users now take for granted: spam rejection, high speed query-based summaries, source clustering, and context(location)-sensitive search. IR and bibliometrics researchers had done it all (relevance, proximity, link analysis, efficiency, scalability, summarization, evaluation) before 1998 but this paper showed how to make it work on the web. For any non-IR engineer attempting to build a web-based retrieval system from scratch, this must be the first port of call."
|9||3/2||Matrix decompositions and latent semantic indexing||Textbook Chapter
18 : Matrix Decompositions and latent semantic indexing
11. "Indexing by latent semantic analysis" by (Deerwester, Dumais, et.al) (*)
Commentary: " IR, as a field, hasn't directly considered the issue of semantic knowledge representation. The above paper is one of the few that does in the following way. LSI is latent semantic analysis (LSA) applied to document retrieval. LSA is actually a variant of a growing ensemble of cognitively-motivated models referred to by the term "semantic space". LSA has an encouraging track record of compatibility with human information processing across a variety of information processing tasks. LSA seems to capture the meaning of words in a way which accords with the representations we carry around in our heads. Finally, the above paper is often cited and interest in LSI seems to have increased markedly in recent years. The above paper has also made an impact outside our field. For example, recent work on latent semantic kernels (machine learning) draws heavily on LSI. "
|10||3/9||Matrix decompositions and latent semantic indexing||Textbook Chapter
8 : Evaluation in Information Retrieval
12. " Unsupervised Named-Entity Extraction from the Web: An Experimental Study " (Etzioni, et.al.) (*)
Commentary: "This paper represents a new generation of IR work that attempts to do more than build a bag of words for information retrieval, but also attempts to make some sense of the information as well."
Exam: no exam
I trust all students are honest and do not cheat. Those who break my trust at any point will get an F in the course - no excuses or apologies will be accepted.Additional penalties may also be imposed by the department and the university. Very severe incidents of academic dishonesty can result in suspension or expulsion from the university.
So don't risk it! If, for some reason, you can't do the homework on time or can't study for the Quiz, you're better off skipping it than cheating it. Do the math!
Students with Disability
Any student who feels he or she may need an accommodation based on the impact of a disability should contact me privately to discuss his or her specific needs. Also contact the Disability Services Center at (949) 824-7494 as soon as possible to better ensure that such accommodations are implemented in a timely fashion.