SlideShare une entreprise Scribd logo
1  sur  23
Efficient Machine Learning and
Machine Learning for Efficiency
in Information Retrieval
Bhaskar Mitra
Principal Researcher
Microsoft Research
@UnderdogGeek bmitra@microsoft.com
Workshop on Reaching Efficiency
in Neural Information Retrieval
Source: https://en.wiktionary.org/wiki/efficient
Complex trade-offs
between different costs
makes benchmarking
challenging
Typical playbook for
efficiency: find cheaper
approximation, trade-off
other costs, reduce task
Machine learning can
be employed to directly
improve efficiency of
large-scale IR systems
Key points:
Challenges of
benchmarking
for efficiency
What do we measure and what do
we control for? How do we capture
trade-offs between measures?
What stages of the pipeline can we
evaluate?
How do we test the benchmark for
construct validity and prevent
negative externalities?
SustaiNLP (2020) Shared Task
Website: https://sites.google.com/view/sustainlp2020/shared-task
Paper: https://aclanthology.org/2020.sustainlp-1.24.pdf
Benchmarks effectiveness and inference-time
energy-consumption
(Some relevant examples from our NLP peers)
NeurIPS 2020 EfficientQA Competition
Website: https://efficientqa.github.io/
Paper: http://proceedings.mlr.press/v133/min21a/min21a.pdf
Benchmarks effectiveness while grouping submitted
systems by their inference-time docker image size
ELUE (Efficient Language Understanding Evaluation)
Website: http://eluebenchmark.fastnlp.top/
Paper: https://txsun1997.github.io/papers/elue_paper.pdf
(Benchmarks effectiveness, number of model
parameters, and inference-time FLOPs)
What would a TREC Neural Efficiency Track look like?
Benchmarking by jointly considering effectiveness,
efficiency, and robustness
Hofstätter, Craswell, Mitra, Zamani, and Hanbury. Are We There Yet? A Decision Framework for Replacing Term-Based Retrieval with Dense Retrieval Systems. ArXiv preprint. (2022)
A preliminary framework:
• Identify key measures of effectiveness, efficiency, and robustness
• Measures can either be traded-off against each other, or act as
guardrails
• Apply value-laden and business-informed trade-offs between
cost and robustness measures to define aggregate measures, e.g.,
• Identify the set of acceptable solutions (again) based on value-
laden and business-informed trade-offs decision boundaries
Developing more efficient neural IR models
This playbook has been useful in allowing explorations of more
ambitious architectures and then reducing their costs /
footprints to make them practically deployable in large
commercial settings
However, we should ask:
• Would we build different models if we are informed by
broader system design (data structures and algorithms) and
cost considerations from the very start?
• Are groups with access to large compute resources
adequately incentivized to consider efficiency as early as
possible in the development process, especially given that
large compute resources can be a competitive advantage?
• Are we missing out on use-cases where machine learning is
employed specifically to improve efficiency, instead of
effectiveness?
Identify key
efficiency
bottlenecks and
costs
Find cheaper
approximations, or
trade-off with other
costs, or reduce the
task
Develop more
complicated
models optimizing
only for
effectiveness
A typical
model
development
playbook
Case studies: Scaling BERT-based relevance models to
long documents and full retrieval settings
Nogueira and Cho (2019)
How do we scale to
longer documents?
How do we scale to
full retrieval from
large collections?
Bottleneck: Peak GPU
memory during training
Challenge: Collection size
vs. expected online query
response time
Challenges in scaling
BERT to longer inputs
At training time, the GPU memory requirement for
BERT’s self-attention layers grows quadratically w.r.t. to
input length
The quadratic complexity is a direct result of storing all
the 𝑛2
-dimensional attention matrices in GPU memory
during training for easier backpropagation
Potential workarounds:
• Trade-off GPU memory and training time by
proactively releasing GPU memory during forward
pass at the cost of redundant re-computations
during backward pass
• Find cheaper approximation to self-attention layers
• Reduce the input space by running BERT on select
passages in the document
Trade-off GPU memory and training time using
gradient checkpointing
At training time, during the forward-pass the model
caches all intermediate outputs in GPU memory so
that during backward-pass we can easily compute
the gradients of a layer’s outputs w.r.t. its inputs
Under gradient checkpointing (Chen et al., 2016), in
contrast, the model only saves intermediate outputs
at specific checkpoints; during the backward-pass,
missing intermediate outputs are recomputed based
on the closed preceding checkpoint(s)
For Transformers, this allows us to store only one
𝑛2
-dimensional attention matrix in GPU memory at
any given time!
Without gradient checkpointing:
With gradient checkpointing: checkpoint
Stored in GPU memory
https://medium.com/tensorflow/fitting-larger-networks-into-memory-583e3c758ff9
Cheaper approximation: Transformer  Conformer
Conformer is an alternative to Transformer that employs a separable self-attention layer with linear GPU
memory complexity (as opposed to Transformer’s quadratic complexity) and is augmented with additional
convolutional layers to model short-distance attention
Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence for Document Retrieval. ArXiv preprint. (2020)
Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence at TREC 2020 Deep Learning Track. In Proc. TREC. (2020)
Mitra, Hofstätter, Zamani, and Craswell. Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence. In Proc. SIGIR. (2021)
TREC 2020 Deep Learning Track
(Document Ranking Task)
Reduce the task: Passage-based document ranking
Hofstätter, Zamani, Mitra, Craswell, and Hanbury. Local Self-Attention over Long Text for Efficient Document Retrieval. In Proc. SIGIR. (2020)
Hofstätter, Mitra, Zamani, Craswell, and Hanbury. Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking. In Proc. SIGIR. (2021)
Kazai, Mitra, Dong, Zamani, Craswell, and Yang. Less is Less: When Are Snippets Insufficient for Human vs Machine Relevance Estimation? In Proc. ECIR. (2022)
Strategy 1: Run BERT on first-k tokens from the document
Considering only the first-k tokens leads to underestimation of relevance and consequently
under-retrieval of longer documents (Hofstätter et al., 2020). Recent studies (Kazai et al., 2022)
have also analyzed when single snippets are insufficient for both human and machine learning
based relevance estimation.
Strategy 2: Run BERT on multiple windows of k-tokens each from the document
This is the approach proposed by Hofstätter et al. (2020). However, the number of windows can be large
corresponding to longer documents and running BERT too many times per query-document pair can also be
prohibitively costly.
Strategy 3: Run BERT on windows of text pre-selected using cheaper models
This is the approach proposed by Hofstätter et al. (2021). The approach
(IDCM) was motivated by cascaded ranking pipelines, but in this case the
cascades are employed within-document for passage selection.
Intra-Document
Cascaded Model (IDCM)
We employ a cascaded architecture: a cheaper model
ranks-and-prunes candidate passages and costlier BERT
model inspects only selected passages from the document
The cheaper model is trained via knowledge distillation
from the BERT model
Hofstätter, Mitra, Zamani, Craswell, and Hanbury. Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking. In Proc. SIGIR. (2021)
Challenges in scaling
BERT to full retrieval
Broadly two sets of approaches have emerged:
Dense retrieval and Query Term Independent
(QTI) models; both precompute document
representations at indexing time and require
very little computations at query response time
Mitra, Rosset, Hawking, Craswell, Diaz, and Yilmaz. Incorporating Query Term Independence Assumption for Efficient Retrieval and Ranking Using Deep Neural Networks. ArXiv preprint. (2019)
Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence for Document Retrieval. ArXiv preprint. (2020)
Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence at TREC 2020 Deep Learning Track. In Proc. TREC. (2020)
Mitra, Hofstätter, Zamani, and Craswell. Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence. In Proc. SIGIR. (2021)
Nogueira and Cho (2019)
Dense retrieval: Xiong et al. (2021), Qu et al.
(2021), Hofstätter et al. (2021), and others
QTI: Mitra et al. (2019), Nogueira et al.
(2019), Dai and Callan (2020), and others
A note about distillation
A popular recipe involves pretraining/finetuning large models and
then knowledge distillation to smaller models that can be deployed in
real-world retrieval systems
Pretrained model
e.g., 24-layer BERT
Finetuned model
e.g., 24-layer BERT
Distilled
model
e.g., smaller
model, or dense
retriever, or QTI,
or early-stage
cascade model
Machine Learning for Retrieval Efficiency
In IR, predictive machine learning has largely been employed for relevance
estimation
Kraska et al. (2018) were one of the earliest to propose learned index structures
where predictive machine learning is employed to speed up search over classical
data structures
Opinion: I believe there’s a significant opportunity to employ deep learning and
other machine learning approaches to directly optimize for efficiency in our
search and recommendation stacks
Let’s look at an example…
Large scale IR systems trade-off search result quality and query response time
In Bing, we have a candidate generation stage followed by multiple rank and prune stages
Typically, we apply machine learning in the re-ranking stages
In this work, we explore reinforcement learning for effective and efficient candidate generation
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
In Bing, the index is distributed over multiple machines
For candidate generation, on each machine the documents are linearly scanned using a match plan
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
When a query comes in, it is automatically
categorized, and a pre-defined match plan
is selected
A match rule defines the condition that a
document should satisfy to be selected as a
candidate
A match plan consists of a sequence of
match rules, and corresponding stopping
criteria
The stopping criteria decides when the
index scan using a particular match rule
should terminate—and if the matching
process should continue with the next match
rule, or conclude, or reset to the beginning
of the index
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
Match plans influence the
trade-off between
effectiveness and efficiency
E.g., long queries with rare
intents may require expensive
match plans that consider
body text and search deeper
into the index
In contrast, for popular
navigational queries a shallow
scan against URL and title
metastreams may be sufficient
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
E.g.,
Query: halloween costumes
Match rule: mrA → (halloween ∈ A|U|B|T ) ∧ (costumes ∈ A|U|B|T )
Query: facebook login
Match rule: mrB → (facebook ∈ U|T )
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
During execution, two accumulators are tracked
u: the number of blocks accessed from disk
v: the cum. number of term matches in all inspected documents
A stopping criteria sets thresholds for each – when either thresholds are met, the scan using
that particular match rule terminates
Matching may then continue with a new match rule, or terminate, or re-start from beginning
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
Optimizing query evaluations using reinforcement learning
Learn a policy πθ : S → A which maximizes the
cumulative discounted reward R, where γ is the
discount rate
We employ table-based Q learning
State space: index blocks accessed (ut) and term
matches (vt)
Action space:
Reward function:
g(di) is the relevance of the ith document estimated based on
the subsequent L1 ranker score—considering only top n
documents
index
match rule relevance discounted by
index blocks accessed
agent
accumulators
(u, v)
Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
In summary…
Complex trade-offs
between different costs
makes benchmarking
challenging
Typical playbook for
efficiency: find cheaper
approximation, trade-off
other costs, reduce task
Machine learning can
be employed to directly
improve efficiency of
large-scale IR systems
How can we create a
shared task to encourage
more efficient deep
learning approaches for IR?
What would deep learning
models for IR look like if
designed with specific
retrieval data-structures in
mind from the start?
What are the key
opportunities to employ
predictive machine
learning to speed up large-
scale retrieval systems?
Questions
to
the
audience…
@UnderdogGeek bmitra@microsoft.com
Thank you!

Contenu connexe

Tendances

Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22GiacomoBalloccu
 
SinGAN - Learning a Generative Model from a Single Natural Image
SinGAN - Learning a Generative Model from a Single Natural ImageSinGAN - Learning a Generative Model from a Single Natural Image
SinGAN - Learning a Generative Model from a Single Natural ImageJishnu P
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Knowledge Graph Embeddings for Recommender Systems
Knowledge Graph Embeddings for Recommender SystemsKnowledge Graph Embeddings for Recommender Systems
Knowledge Graph Embeddings for Recommender SystemsEnrico Palumbo
 
Implications of GPT-3
Implications of GPT-3Implications of GPT-3
Implications of GPT-3Raven Jiang
 
Yurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAI
Yurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAIYurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAI
Yurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAILviv Startup Club
 
Fundamentals of Deep Recommender Systems
 Fundamentals of Deep Recommender Systems Fundamentals of Deep Recommender Systems
Fundamentals of Deep Recommender SystemsWQ Fan
 
Java Interview Questions by NageswaraRao
Java Interview Questions by NageswaraRaoJava Interview Questions by NageswaraRao
Java Interview Questions by NageswaraRaoJavabynataraJ
 
Book Recommendation System using Data Mining for the University of Hong Kong ...
Book Recommendation System using Data Mining for the University of Hong Kong ...Book Recommendation System using Data Mining for the University of Hong Kong ...
Book Recommendation System using Data Mining for the University of Hong Kong ...CITE
 
Scaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language ModelsScaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language Modelstaeseon ryu
 
Glove global vectors for word representation
Glove global vectors for word representationGlove global vectors for word representation
Glove global vectors for word representationhyunyoung Lee
 
Neural Language Generation Head to Toe
Neural Language Generation Head to Toe Neural Language Generation Head to Toe
Neural Language Generation Head to Toe Hady Elsahar
 
Generative Adversarial Networks
Generative Adversarial NetworksGenerative Adversarial Networks
Generative Adversarial NetworksMark Chang
 
Generative Adversarial Networks and Their Applications in Medical Imaging
Generative Adversarial Networks  and Their Applications in Medical ImagingGenerative Adversarial Networks  and Their Applications in Medical Imaging
Generative Adversarial Networks and Their Applications in Medical ImagingSanghoon Hong
 
Boston ML - Architecting Recommender Systems
Boston ML - Architecting Recommender SystemsBoston ML - Architecting Recommender Systems
Boston ML - Architecting Recommender SystemsJames Kirk
 
Deep Belief Networks
Deep Belief NetworksDeep Belief Networks
Deep Belief NetworksHasan H Topcu
 

Tendances (20)

Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
 
SinGAN - Learning a Generative Model from a Single Natural Image
SinGAN - Learning a Generative Model from a Single Natural ImageSinGAN - Learning a Generative Model from a Single Natural Image
SinGAN - Learning a Generative Model from a Single Natural Image
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Knowledge Graph Embeddings for Recommender Systems
Knowledge Graph Embeddings for Recommender SystemsKnowledge Graph Embeddings for Recommender Systems
Knowledge Graph Embeddings for Recommender Systems
 
Implications of GPT-3
Implications of GPT-3Implications of GPT-3
Implications of GPT-3
 
Java beans
Java beansJava beans
Java beans
 
Yurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAI
Yurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAIYurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAI
Yurii Pashchenko: Zero-shot learning capabilities of CLIP model from OpenAI
 
Fundamentals of Deep Recommender Systems
 Fundamentals of Deep Recommender Systems Fundamentals of Deep Recommender Systems
Fundamentals of Deep Recommender Systems
 
Java Interview Questions by NageswaraRao
Java Interview Questions by NageswaraRaoJava Interview Questions by NageswaraRao
Java Interview Questions by NageswaraRao
 
Book Recommendation System using Data Mining for the University of Hong Kong ...
Book Recommendation System using Data Mining for the University of Hong Kong ...Book Recommendation System using Data Mining for the University of Hong Kong ...
Book Recommendation System using Data Mining for the University of Hong Kong ...
 
Yapay Zeka ve Makine Öğrenmesi
Yapay Zeka ve Makine ÖğrenmesiYapay Zeka ve Makine Öğrenmesi
Yapay Zeka ve Makine Öğrenmesi
 
Java exception-handling
Java exception-handlingJava exception-handling
Java exception-handling
 
Scaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language ModelsScaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language Models
 
Glove global vectors for word representation
Glove global vectors for word representationGlove global vectors for word representation
Glove global vectors for word representation
 
Neural Language Generation Head to Toe
Neural Language Generation Head to Toe Neural Language Generation Head to Toe
Neural Language Generation Head to Toe
 
Generative Adversarial Networks
Generative Adversarial NetworksGenerative Adversarial Networks
Generative Adversarial Networks
 
Generative Adversarial Networks and Their Applications in Medical Imaging
Generative Adversarial Networks  and Their Applications in Medical ImagingGenerative Adversarial Networks  and Their Applications in Medical Imaging
Generative Adversarial Networks and Their Applications in Medical Imaging
 
Boston ML - Architecting Recommender Systems
Boston ML - Architecting Recommender SystemsBoston ML - Architecting Recommender Systems
Boston ML - Architecting Recommender Systems
 
Recommender Systems
Recommender SystemsRecommender Systems
Recommender Systems
 
Deep Belief Networks
Deep Belief NetworksDeep Belief Networks
Deep Belief Networks
 

Similaire à Efficient Machine Learning and Machine Learning for Efficiency in Information Retrieval (Keynote @ SIGIR 2022 ReNeuIR Workshop)

Multi-threaded approach in generating frequent itemset of Apriori algorithm b...
Multi-threaded approach in generating frequent itemset of Apriori algorithm b...Multi-threaded approach in generating frequent itemset of Apriori algorithm b...
Multi-threaded approach in generating frequent itemset of Apriori algorithm b...TELKOMNIKA JOURNAL
 
OpenACC and Hackathons Monthly Highlights: April 2023
OpenACC and Hackathons Monthly Highlights: April  2023OpenACC and Hackathons Monthly Highlights: April  2023
OpenACC and Hackathons Monthly Highlights: April 2023OpenACC
 
What deep learning can bring to...
What deep learning can bring to...What deep learning can bring to...
What deep learning can bring to...Gautier Marti
 
Bäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdf
Bäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdfBäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdf
Bäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdfTadiyosHailemichael
 
Agile performance engineering with cloud 2016
Agile performance engineering with cloud   2016Agile performance engineering with cloud   2016
Agile performance engineering with cloud 2016Ken Chan
 
Functional Verification of Large-integers Circuits using a Cosimulation-base...
Functional Verification of Large-integers Circuits using a  Cosimulation-base...Functional Verification of Large-integers Circuits using a  Cosimulation-base...
Functional Verification of Large-integers Circuits using a Cosimulation-base...IJECEIAES
 
MultiObjective(11) - Copy
MultiObjective(11) - CopyMultiObjective(11) - Copy
MultiObjective(11) - CopyAMIT KUMAR
 
IRJET- Review of Existing Methods in K-Means Clustering Algorithm
IRJET- Review of Existing Methods in K-Means Clustering AlgorithmIRJET- Review of Existing Methods in K-Means Clustering Algorithm
IRJET- Review of Existing Methods in K-Means Clustering AlgorithmIRJET Journal
 
Performance Comparison between Pytorch and Mindspore
Performance Comparison between Pytorch and MindsporePerformance Comparison between Pytorch and Mindspore
Performance Comparison between Pytorch and Mindsporeijdms
 
A REVIEW ON PARALLEL COMPUTING
A REVIEW ON PARALLEL COMPUTINGA REVIEW ON PARALLEL COMPUTING
A REVIEW ON PARALLEL COMPUTINGAmy Roman
 
IRJET- Diverse Approaches for Document Clustering in Product Development Anal...
IRJET- Diverse Approaches for Document Clustering in Product Development Anal...IRJET- Diverse Approaches for Document Clustering in Product Development Anal...
IRJET- Diverse Approaches for Document Clustering in Product Development Anal...IRJET Journal
 
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...acijjournal
 
Deepcoder to Self-Code with Machine Learning
Deepcoder to Self-Code with Machine LearningDeepcoder to Self-Code with Machine Learning
Deepcoder to Self-Code with Machine LearningIRJET Journal
 
ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA
 ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA
ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATANexgen Technology
 
Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...
Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...
Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...Facultad de Informática UCM
 
Collective Mind: a collaborative curation tool for program optimization
Collective Mind: a collaborative curation tool for program optimizationCollective Mind: a collaborative curation tool for program optimization
Collective Mind: a collaborative curation tool for program optimizationGrigori Fursin
 
2023comp90024_Spartan.pdf
2023comp90024_Spartan.pdf2023comp90024_Spartan.pdf
2023comp90024_Spartan.pdfLevLafayette1
 
IEEE Parallel and distributed system 2016 Title and Abstract
IEEE Parallel and distributed system 2016 Title and AbstractIEEE Parallel and distributed system 2016 Title and Abstract
IEEE Parallel and distributed system 2016 Title and Abstracttsysglobalsolutions
 
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...IAEME Publication
 
2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )
2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )
2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )SBGC
 

Similaire à Efficient Machine Learning and Machine Learning for Efficiency in Information Retrieval (Keynote @ SIGIR 2022 ReNeuIR Workshop) (20)

Multi-threaded approach in generating frequent itemset of Apriori algorithm b...
Multi-threaded approach in generating frequent itemset of Apriori algorithm b...Multi-threaded approach in generating frequent itemset of Apriori algorithm b...
Multi-threaded approach in generating frequent itemset of Apriori algorithm b...
 
OpenACC and Hackathons Monthly Highlights: April 2023
OpenACC and Hackathons Monthly Highlights: April  2023OpenACC and Hackathons Monthly Highlights: April  2023
OpenACC and Hackathons Monthly Highlights: April 2023
 
What deep learning can bring to...
What deep learning can bring to...What deep learning can bring to...
What deep learning can bring to...
 
Bäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdf
Bäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdfBäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdf
Bäßler2022_Article_UnsupervisedAnomalyDetectionIn.pdf
 
Agile performance engineering with cloud 2016
Agile performance engineering with cloud   2016Agile performance engineering with cloud   2016
Agile performance engineering with cloud 2016
 
Functional Verification of Large-integers Circuits using a Cosimulation-base...
Functional Verification of Large-integers Circuits using a  Cosimulation-base...Functional Verification of Large-integers Circuits using a  Cosimulation-base...
Functional Verification of Large-integers Circuits using a Cosimulation-base...
 
MultiObjective(11) - Copy
MultiObjective(11) - CopyMultiObjective(11) - Copy
MultiObjective(11) - Copy
 
IRJET- Review of Existing Methods in K-Means Clustering Algorithm
IRJET- Review of Existing Methods in K-Means Clustering AlgorithmIRJET- Review of Existing Methods in K-Means Clustering Algorithm
IRJET- Review of Existing Methods in K-Means Clustering Algorithm
 
Performance Comparison between Pytorch and Mindspore
Performance Comparison between Pytorch and MindsporePerformance Comparison between Pytorch and Mindspore
Performance Comparison between Pytorch and Mindspore
 
A REVIEW ON PARALLEL COMPUTING
A REVIEW ON PARALLEL COMPUTINGA REVIEW ON PARALLEL COMPUTING
A REVIEW ON PARALLEL COMPUTING
 
IRJET- Diverse Approaches for Document Clustering in Product Development Anal...
IRJET- Diverse Approaches for Document Clustering in Product Development Anal...IRJET- Diverse Approaches for Document Clustering in Product Development Anal...
IRJET- Diverse Approaches for Document Clustering in Product Development Anal...
 
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...
 
Deepcoder to Self-Code with Machine Learning
Deepcoder to Self-Code with Machine LearningDeepcoder to Self-Code with Machine Learning
Deepcoder to Self-Code with Machine Learning
 
ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA
 ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA
ON DISTRIBUTED FUZZY DECISION TREES FOR BIG DATA
 
Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...
Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...
Fast and energy-efficient eNVM based memory organisation at L3-L1 layers for ...
 
Collective Mind: a collaborative curation tool for program optimization
Collective Mind: a collaborative curation tool for program optimizationCollective Mind: a collaborative curation tool for program optimization
Collective Mind: a collaborative curation tool for program optimization
 
2023comp90024_Spartan.pdf
2023comp90024_Spartan.pdf2023comp90024_Spartan.pdf
2023comp90024_Spartan.pdf
 
IEEE Parallel and distributed system 2016 Title and Abstract
IEEE Parallel and distributed system 2016 Title and AbstractIEEE Parallel and distributed system 2016 Title and Abstract
IEEE Parallel and distributed system 2016 Title and Abstract
 
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...
 
2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )
2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )
2017 IEEE Projects 2017 For Cse ( Trichy, Chennai )
 

Plus de Bhaskar Mitra

Neural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progressNeural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progressBhaskar Mitra
 
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackConformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Duet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackDuet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackBhaskar Mitra
 
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBenchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBhaskar Mitra
 
Deep Neural Methods for Retrieval
Deep Neural Methods for RetrievalDeep Neural Methods for Retrieval
Deep Neural Methods for RetrievalBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Learning to Rank with Neural Networks
Learning to Rank with Neural NetworksLearning to Rank with Neural Networks
Learning to Rank with Neural NetworksBhaskar Mitra
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for SearchBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for SearchBhaskar Mitra
 
Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Bhaskar Mitra
 
Adversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrievalAdversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrievalBhaskar Mitra
 
5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information RetrievalBhaskar Mitra
 
A Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information RetrievalA Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information RetrievalBhaskar Mitra
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information RetrievalBhaskar Mitra
 
Neural Models for Document Ranking
Neural Models for Document RankingNeural Models for Document Ranking
Neural Models for Document RankingBhaskar Mitra
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information RetrievalBhaskar Mitra
 
Neu-IR 2017: welcome
Neu-IR 2017: welcomeNeu-IR 2017: welcome
Neu-IR 2017: welcomeBhaskar Mitra
 

Plus de Bhaskar Mitra (20)

Neural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progressNeural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progress
 
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackConformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Duet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackDuet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning Track
 
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBenchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
 
Deep Neural Methods for Retrieval
Deep Neural Methods for RetrievalDeep Neural Methods for Retrieval
Deep Neural Methods for Retrieval
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Learning to Rank with Neural Networks
Learning to Rank with Neural NetworksLearning to Rank with Neural Networks
Learning to Rank with Neural Networks
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for Search
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for Search
 
Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)
 
Adversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrievalAdversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrieval
 
5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval
 
A Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information RetrievalA Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information Retrieval
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information Retrieval
 
Neural Models for Document Ranking
Neural Models for Document RankingNeural Models for Document Ranking
Neural Models for Document Ranking
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information Retrieval
 
Neu-IR 2017: welcome
Neu-IR 2017: welcomeNeu-IR 2017: welcome
Neu-IR 2017: welcome
 
The Duet model
The Duet modelThe Duet model
The Duet model
 

Dernier

TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...apidays
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Jeffrey Haguewood
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdfSandro Moreira
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingEdi Saputra
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MIND CTI
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...Zilliz
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...apidays
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodJuan lago vázquez
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Angeliki Cooney
 
Cyberprint. Dark Pink Apt Group [EN].pdf
Cyberprint. Dark Pink Apt Group [EN].pdfCyberprint. Dark Pink Apt Group [EN].pdf
Cyberprint. Dark Pink Apt Group [EN].pdfOverkill Security
 
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Zilliz
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusZilliz
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyKhushali Kathiriya
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAndrey Devyatkin
 

Dernier (20)

TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
 
Cyberprint. Dark Pink Apt Group [EN].pdf
Cyberprint. Dark Pink Apt Group [EN].pdfCyberprint. Dark Pink Apt Group [EN].pdf
Cyberprint. Dark Pink Apt Group [EN].pdf
 
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 

Efficient Machine Learning and Machine Learning for Efficiency in Information Retrieval (Keynote @ SIGIR 2022 ReNeuIR Workshop)

  • 1. Efficient Machine Learning and Machine Learning for Efficiency in Information Retrieval Bhaskar Mitra Principal Researcher Microsoft Research @UnderdogGeek bmitra@microsoft.com Workshop on Reaching Efficiency in Neural Information Retrieval
  • 3. Complex trade-offs between different costs makes benchmarking challenging Typical playbook for efficiency: find cheaper approximation, trade-off other costs, reduce task Machine learning can be employed to directly improve efficiency of large-scale IR systems Key points:
  • 4. Challenges of benchmarking for efficiency What do we measure and what do we control for? How do we capture trade-offs between measures? What stages of the pipeline can we evaluate? How do we test the benchmark for construct validity and prevent negative externalities? SustaiNLP (2020) Shared Task Website: https://sites.google.com/view/sustainlp2020/shared-task Paper: https://aclanthology.org/2020.sustainlp-1.24.pdf Benchmarks effectiveness and inference-time energy-consumption (Some relevant examples from our NLP peers) NeurIPS 2020 EfficientQA Competition Website: https://efficientqa.github.io/ Paper: http://proceedings.mlr.press/v133/min21a/min21a.pdf Benchmarks effectiveness while grouping submitted systems by their inference-time docker image size ELUE (Efficient Language Understanding Evaluation) Website: http://eluebenchmark.fastnlp.top/ Paper: https://txsun1997.github.io/papers/elue_paper.pdf (Benchmarks effectiveness, number of model parameters, and inference-time FLOPs) What would a TREC Neural Efficiency Track look like?
  • 5. Benchmarking by jointly considering effectiveness, efficiency, and robustness Hofstätter, Craswell, Mitra, Zamani, and Hanbury. Are We There Yet? A Decision Framework for Replacing Term-Based Retrieval with Dense Retrieval Systems. ArXiv preprint. (2022) A preliminary framework: • Identify key measures of effectiveness, efficiency, and robustness • Measures can either be traded-off against each other, or act as guardrails • Apply value-laden and business-informed trade-offs between cost and robustness measures to define aggregate measures, e.g., • Identify the set of acceptable solutions (again) based on value- laden and business-informed trade-offs decision boundaries
  • 6. Developing more efficient neural IR models This playbook has been useful in allowing explorations of more ambitious architectures and then reducing their costs / footprints to make them practically deployable in large commercial settings However, we should ask: • Would we build different models if we are informed by broader system design (data structures and algorithms) and cost considerations from the very start? • Are groups with access to large compute resources adequately incentivized to consider efficiency as early as possible in the development process, especially given that large compute resources can be a competitive advantage? • Are we missing out on use-cases where machine learning is employed specifically to improve efficiency, instead of effectiveness? Identify key efficiency bottlenecks and costs Find cheaper approximations, or trade-off with other costs, or reduce the task Develop more complicated models optimizing only for effectiveness A typical model development playbook
  • 7. Case studies: Scaling BERT-based relevance models to long documents and full retrieval settings Nogueira and Cho (2019) How do we scale to longer documents? How do we scale to full retrieval from large collections? Bottleneck: Peak GPU memory during training Challenge: Collection size vs. expected online query response time
  • 8. Challenges in scaling BERT to longer inputs At training time, the GPU memory requirement for BERT’s self-attention layers grows quadratically w.r.t. to input length The quadratic complexity is a direct result of storing all the 𝑛2 -dimensional attention matrices in GPU memory during training for easier backpropagation Potential workarounds: • Trade-off GPU memory and training time by proactively releasing GPU memory during forward pass at the cost of redundant re-computations during backward pass • Find cheaper approximation to self-attention layers • Reduce the input space by running BERT on select passages in the document
  • 9. Trade-off GPU memory and training time using gradient checkpointing At training time, during the forward-pass the model caches all intermediate outputs in GPU memory so that during backward-pass we can easily compute the gradients of a layer’s outputs w.r.t. its inputs Under gradient checkpointing (Chen et al., 2016), in contrast, the model only saves intermediate outputs at specific checkpoints; during the backward-pass, missing intermediate outputs are recomputed based on the closed preceding checkpoint(s) For Transformers, this allows us to store only one 𝑛2 -dimensional attention matrix in GPU memory at any given time! Without gradient checkpointing: With gradient checkpointing: checkpoint Stored in GPU memory https://medium.com/tensorflow/fitting-larger-networks-into-memory-583e3c758ff9
  • 10. Cheaper approximation: Transformer  Conformer Conformer is an alternative to Transformer that employs a separable self-attention layer with linear GPU memory complexity (as opposed to Transformer’s quadratic complexity) and is augmented with additional convolutional layers to model short-distance attention Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence for Document Retrieval. ArXiv preprint. (2020) Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence at TREC 2020 Deep Learning Track. In Proc. TREC. (2020) Mitra, Hofstätter, Zamani, and Craswell. Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence. In Proc. SIGIR. (2021) TREC 2020 Deep Learning Track (Document Ranking Task)
  • 11. Reduce the task: Passage-based document ranking Hofstätter, Zamani, Mitra, Craswell, and Hanbury. Local Self-Attention over Long Text for Efficient Document Retrieval. In Proc. SIGIR. (2020) Hofstätter, Mitra, Zamani, Craswell, and Hanbury. Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking. In Proc. SIGIR. (2021) Kazai, Mitra, Dong, Zamani, Craswell, and Yang. Less is Less: When Are Snippets Insufficient for Human vs Machine Relevance Estimation? In Proc. ECIR. (2022) Strategy 1: Run BERT on first-k tokens from the document Considering only the first-k tokens leads to underestimation of relevance and consequently under-retrieval of longer documents (Hofstätter et al., 2020). Recent studies (Kazai et al., 2022) have also analyzed when single snippets are insufficient for both human and machine learning based relevance estimation. Strategy 2: Run BERT on multiple windows of k-tokens each from the document This is the approach proposed by Hofstätter et al. (2020). However, the number of windows can be large corresponding to longer documents and running BERT too many times per query-document pair can also be prohibitively costly. Strategy 3: Run BERT on windows of text pre-selected using cheaper models This is the approach proposed by Hofstätter et al. (2021). The approach (IDCM) was motivated by cascaded ranking pipelines, but in this case the cascades are employed within-document for passage selection.
  • 12. Intra-Document Cascaded Model (IDCM) We employ a cascaded architecture: a cheaper model ranks-and-prunes candidate passages and costlier BERT model inspects only selected passages from the document The cheaper model is trained via knowledge distillation from the BERT model Hofstätter, Mitra, Zamani, Craswell, and Hanbury. Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking. In Proc. SIGIR. (2021)
  • 13. Challenges in scaling BERT to full retrieval Broadly two sets of approaches have emerged: Dense retrieval and Query Term Independent (QTI) models; both precompute document representations at indexing time and require very little computations at query response time Mitra, Rosset, Hawking, Craswell, Diaz, and Yilmaz. Incorporating Query Term Independence Assumption for Efficient Retrieval and Ranking Using Deep Neural Networks. ArXiv preprint. (2019) Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence for Document Retrieval. ArXiv preprint. (2020) Mitra, Hofstätter, Zamani, and Craswell. Conformer-Kernel with Query Term Independence at TREC 2020 Deep Learning Track. In Proc. TREC. (2020) Mitra, Hofstätter, Zamani, and Craswell. Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence. In Proc. SIGIR. (2021) Nogueira and Cho (2019) Dense retrieval: Xiong et al. (2021), Qu et al. (2021), Hofstätter et al. (2021), and others QTI: Mitra et al. (2019), Nogueira et al. (2019), Dai and Callan (2020), and others
  • 14. A note about distillation A popular recipe involves pretraining/finetuning large models and then knowledge distillation to smaller models that can be deployed in real-world retrieval systems Pretrained model e.g., 24-layer BERT Finetuned model e.g., 24-layer BERT Distilled model e.g., smaller model, or dense retriever, or QTI, or early-stage cascade model
  • 15. Machine Learning for Retrieval Efficiency In IR, predictive machine learning has largely been employed for relevance estimation Kraska et al. (2018) were one of the earliest to propose learned index structures where predictive machine learning is employed to speed up search over classical data structures Opinion: I believe there’s a significant opportunity to employ deep learning and other machine learning approaches to directly optimize for efficiency in our search and recommendation stacks Let’s look at an example…
  • 16. Large scale IR systems trade-off search result quality and query response time In Bing, we have a candidate generation stage followed by multiple rank and prune stages Typically, we apply machine learning in the re-ranking stages In this work, we explore reinforcement learning for effective and efficient candidate generation Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 17. In Bing, the index is distributed over multiple machines For candidate generation, on each machine the documents are linearly scanned using a match plan Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 18. When a query comes in, it is automatically categorized, and a pre-defined match plan is selected A match rule defines the condition that a document should satisfy to be selected as a candidate A match plan consists of a sequence of match rules, and corresponding stopping criteria The stopping criteria decides when the index scan using a particular match rule should terminate—and if the matching process should continue with the next match rule, or conclude, or reset to the beginning of the index Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 19. Match plans influence the trade-off between effectiveness and efficiency E.g., long queries with rare intents may require expensive match plans that consider body text and search deeper into the index In contrast, for popular navigational queries a shallow scan against URL and title metastreams may be sufficient Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 20. E.g., Query: halloween costumes Match rule: mrA → (halloween ∈ A|U|B|T ) ∧ (costumes ∈ A|U|B|T ) Query: facebook login Match rule: mrB → (facebook ∈ U|T ) Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 21. During execution, two accumulators are tracked u: the number of blocks accessed from disk v: the cum. number of term matches in all inspected documents A stopping criteria sets thresholds for each – when either thresholds are met, the scan using that particular match rule terminates Matching may then continue with a new match rule, or terminate, or re-start from beginning Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 22. Optimizing query evaluations using reinforcement learning Learn a policy πθ : S → A which maximizes the cumulative discounted reward R, where γ is the discount rate We employ table-based Q learning State space: index blocks accessed (ut) and term matches (vt) Action space: Reward function: g(di) is the relevance of the ith document estimated based on the subsequent L1 ranker score—considering only top n documents index match rule relevance discounted by index blocks accessed agent accumulators (u, v) Rosset, Jose, Ghosh, Mitra, and Tiwary. Optimizing Query Evaluations Using Reinforcement Learning for Web Search. In Proc. SIGIR. (2018)
  • 23. In summary… Complex trade-offs between different costs makes benchmarking challenging Typical playbook for efficiency: find cheaper approximation, trade-off other costs, reduce task Machine learning can be employed to directly improve efficiency of large-scale IR systems How can we create a shared task to encourage more efficient deep learning approaches for IR? What would deep learning models for IR look like if designed with specific retrieval data-structures in mind from the start? What are the key opportunities to employ predictive machine learning to speed up large- scale retrieval systems? Questions to the audience… @UnderdogGeek bmitra@microsoft.com Thank you!