→ rod24574575 : 收錄資訊系精華區! 12/30 22:55
課程名稱︰資訊檢索與擷取
課程性質︰資工系選修
課程教師︰陳信希
開課學院:電機資訊學院
開課系所︰資訊工程學系
考試日期(年月日)︰2022/01/06
考試時限(分鐘):180
試題 :
1. The following lists 5 tasks and 10 evaluation metrics.
Tasks: Web Search, Question Answering, Named Entity Recognition, Relation Ex-
traction, Entity Retrieval
Evaluation Metrics: Accuracy, Precision, Recall, F1, F0.5, F2, MAP, NDCG,
MRR, Kendall Tau Coefficient
(a) Please discuss and explain which evaluation metrics are suitable for each
task. If there are no suitable metrics for a task, please give your sug-
gestions. (25 points)
(b) If there are no suitable tasks for an evaluation metric, please discuss
in what situation the evaluation metric will be adopted. (10 points)
2. Long documents may contain mixture of topics. Query matches may be spread
over the whole document. Please describe how a neural document ranking model
aggregates the relevant metches from different parts of a long document. (10
points)
3. Knowledge Base Acceleration (KBA) task is defined as follows.
This task aims to filter a time-ordered corpus for documents that are highly
relevant to a predefined list of entities. Total 27 people and 2 organiza-
tions are selected. A stream corpus spanning 4,973 consecutive hours is con-
structed. It contains over 400M documents. Each document has a timestamp that
places it in the stream. The 29 target entities were mentioned infrequently
enough in the corpus. Judgments for documents from before stream corpus con-
struction time were provided as training data for filtering documents from
the remaining hours. You are instructed to apply your system to each hourly
directory of corpus data in chronological order. For each hour, before pro-
cessing the next hour, systems are expected to emit a list of assertions con-
necting documents and entities. The goal is to identify only central-rated
documents.
(a) Please show your idea to deal with the KBA task. (10 points)
(b) Please discuss how this task is related to Knowledge Base Completion
(KBC), which involves in discovering missing facts. (5 points)
4. Knowledge base is useful for document retrieval. Please explain the latent
factor modeling approach and the deep learning approach to introduce know-
ledge base to enhance the performance of document retrieval. (14 points)
5. Entity relationship explanation is a textual description to describe how a
given pair of entities is related. Please show how to deal with this task by
using knowledge graph. (6 points)
6. Traditional IE predict their relation from a predefined set such as "Birth-
Place" and "Spouse." By contrast, open information extraction (Open IE) aims
to extract the triples that consist of a pair of argument phrases and their
relation phrase from textual data. For example, one can extract the following
two triples from the sentence "Albert Einstein was born in Ulm and died in
Princeton."
(Albert Einstein, was born in, Ulm)
(Albert Einstein, died in, Princeton)
Please answer the following questions about open IE.
(a) Compared with traditional IE, give an advantage and a disadvantage of
open IE. (6 points)
(b) Give two downstream applications of open IE. (6 points)
(c) Given a collection of news articles, please provide a feasible method to
construct an open IE system without the need of labeled data. (8 points)
7. The following lists the presentation topics presented by the team members.
Team 1:
Learning an End-to-End Structure for Retrieval in Large-Scale Recommendations
Team 2:
1. EmbedKGQA: Improving Multi-hop Question Answering over Knowledge Graphs
using Knowledge Base Embeddings
2. TransferNet: An Effective and Transparent Framework for Multi-hop Question
Answering over Relation Graph
3. Improving Multi-hop Knowledge Base Question Answering by Learning Interme-
diate Supervision Signals
Team 3:
Inductive Topic Variational Graph Auto-Encoder for Text Classification
Team 4:
Dense Passage Retrieval for Open-Domain Question Answering
Team 5:
"Did you buy it already?", Detecting Users Purchase-State From Their Product-
Related Questions
Team 6:
UnitedQA: A Hybrid Approach for Open Domain Question Answering
Team 7:
1. A Reinforcement Learning Framework for Relevance Feedback
2. Generating Images Instead of Retrieving Them: Relevance Feedback on Gene-
rative Adversarial Networks
Team 8:
1. AutoDebias: Learning to Debias for Recommendation
2. Casual Intervention for Leveraging Popularity Bias in Recommendation
Team 9:
Self-Supervised Reinforcement Learning for Recommender Systems
Team 10:
1. Multi-behavior Recommendation with Graph Convolutional Networks.
2. Graph Heterogeneous Multi-Relational Recommendation.
Team 11:
Self-supervised Graph Learning for Recommentation.
Team 12:
Personalized Search-based Query Rewrite System for Conversational AI
Team 13:
Group based Personalized Search by Integrating Search Behaviour and Friend
Network
Team 14:
Answering Any-hop Open-domain Questions with Iterative Document Reranking
Team 15:
1. Time Matters: Sequential Recommendation with Complex Temporal Information
2. Motif-aware Sequential Recommendation
Team 16:
1. Recommending Podcasts for Cold-Start Users Based on Music Listening and
Taste
2. Fairness among New Items in Cold Start Recommender Systems
3. A Heterogeneous Graph Neural Model for Cold-start Recommendation
Team 17:
1. Estimation-Action-Reflection: Towards Deep Interaction Between Conversa-
tional and Recommender Systems
2. Time Interval Aware Self-Attention for Sequential Recommendation
Please write down your team id first, and then select the most exciting topic
you learned from the other team. Please write down this team id and specify
the idea you learned from their presentation in brief. (10 points)
--
第01話 似乎在課堂上聽過的樣子 第02話 那真是太令人絕望了
第03話 已經沒什麼好期望了 第04話 被當、21都是存在的
第05話 怎麼可能會all pass 第06話 這考卷絕對有問題啊
第07話 你能面對真正的分數嗎 第08話 我,真是個笨蛋
第09話 這樣成績,教授絕不會讓我過的 第10話 再也不依靠考古題
第11話 最後留下的補考 第12話 我最愛的學分
--
※ 發信站: 批踢踢實業坊(ptt.cc), 來自: 111.249.65.236 (臺灣)
※ 文章網址: https://www.ptt.cc/bbs/NTU-Exam/M.1767060936.A.BB4.html