How is mobile technology and search engine tuning evolving to meet the needs of users? Here we look at recent developments in research, implementations by search engines, and how to look at reach users can adapt their strategies to take into account these next-level changes.
34. 'The Vocabulary
Problem’
Furnas, G.W., Landauer, T.K.,
Gomez, L.M. and Dumais,
S.T., 1987. The vocabulary
problem in human-system
communication. Communicatio
ns of the ACM, 30(11), pp.964-
971.
1987
35.
36. One of the inventors of
‘Latent Semantic
Indexing’, created to
solve ‘The Vocabulary
Problem’ whilst
researching at Bellcore
(1990)
48. Like RPG (Role Playing Games)
Where choices made determine the next choices given
49. People also ask / related queries are a special kind of ‘Query
Refinement’
50. A kind of
‘probability-driven
fork in the road’
http://delivery.acm.org/10.1145/1780000/1772776/p841-
sadikov.pdf (Sadikov et al, 2010) CLUSTERING
QUERY REFINEMENTS BY USER INTENT
51. BUT word’s meaning &
user intent /context
combined are still very
hard to understand for
search engines
52. Despite huge leaps forward in
query classification & natural
language understanding
92. “Easier if we can model: who is
asking, what they have done in
the past, where they are, when it
is, etc.” (Susan Dumais, CIKM,
2016)
93. “Queries Are Difficult To Understand
in Isolation” (Susan Dumais,
Microsoft Research, 2016)
94. AKA - Contextual Search =
User + Time + Location + Device
+ Task
95. Better still… what about predicting
the user’s informational needs to
proactively make suggestions
96. “Nevertheless, as the world is
becoming more mobile-centric, this
old-fashioned query-driven search
scenario and clickbased evaluation
mechanism can no longer catch up with
the rapid evolution of user demand on
mobile devices.” (Song and Guo,2016
(Microsoft Research))
99. Personalising Search via Interests & Activities
2005 paper awarded the 2017 SIGIR Test of Time Award. Cited 1029 times to date
Teevan, J., Dumais, S.T. and Horvitz, E., 2005, August. Personalizing search via automated
analysis of interests and activities. In Proceedings of the 28th annual international ACM
143. Understand your customers to assist with AI
Perceived
Information need
Micro-task
Micro-task Micro-task Micro-task Micro-task Task
Micro-task Micro-task Micro-task Micro-task Task
Micro-task Micro-task Task
Micro-task
Micro-task
Micro-task Task
Micro-task Micro-task Task
Micro-task Task
We can identify the user’s
probable top tasks &
subtasks
Identify their needs & what
info they need along the
way
144. Tell us about
the tasks, order
and steps
involved in
booking a hotel
176. Go
• Go big on evergreen content & keep updated
Optimise
• Optimise images well – think curation / collections
Map
• Map user journeys to content plans
Optimise
• video well – enhance with markup / transcription
Get
• Get personal – keep refining segments / personas
Identify
• Identify & cluster content around task timelines
Use
• Use relatedness across content, tasks & temporality
180. You may need a
dual or multi-
armed content
strategy
2%
181. Book hotel
intent
When do you
want to stay?
dates
dates
How many
nights?
3 nights 2 nights
Overnight A week
Single or
double room?
Single room Double room
Programme your
own expected
questions and
answers
182. References
• Broder, A., 2002, September. A taxonomy of web search. In ACM Sigir forum (Vol. 36, No.
2, pp. 3-10). ACM.
• Chuklin, A., Severyn, A., Trippas, J., Alfonseca, E., Silen, H. and Spina, D., 2018. Prosody
Modifications for Question-Answering in Voice-Only Settings. arXiv preprint
arXiv:1806.03957.
• HigherVisibility. 2018. How Popular is Voice Search? | HigherVisibility. [ONLINE] Available
at: https://www.highervisibility.com/blog/how-popular-is-voice-search/
• Filippova, K., Alfonseca, E., Colmenares, C.A., Kaiser, L. and Vinyals, O., 2015. Sentence
compression by deletion with lstms. In Proceedings of the 2015 Conference on Empirical
Methods in Natural Language Processing (pp. 360-368).
• Filippova, K. and Alfonseca, E., 2015. Fast k-best sentence compression. arXiv preprint
arXiv:1510.08418.
• Google Developers. 2018. Content-based Actions | Actions on Google | Google
Developers. [ONLINE] Available at: https://developers.google.com/actions/content-
actions/. [Accessed 18 June 2018]
183. References
• Mitkov, R., 2014. Anaphora resolution. Routledge.
• NLP Department - Stanford University - Imran Q Sayed. 2018. Issues in Anaphora Resolution.
[ONLINE] Available
at: https://nlp.stanford.edu/courses/cs224n/2003/fp/iqsayed/project_report.pdf. [Accessed 28
June 2018].
• Radlinski, F. and Craswell, N., 2017, March. A theoretical framework for conversational search.
In Proceedings of the 2017 Conference on Conference Human Information Interaction and
Retrieval (pp. 117-126). ACM.
• Schalkwyk, J., Beeferman, D., Beaufays, F., Byrne, B., Chelba, C., Cohen, M., Kamvar, M. and
Strope, B., 2010. “Your word is my command”: Google search by voice: a case study. In Advances
in speech recognition (pp. 61-90). Springer, Boston, MA.
• SISTRIX. 2018. Stepping out of the SEO Bubble - SISTRIX. [ONLINE] Available
at: https://www.sistrix.com/blog/stepping-out-of-the-seo-bubble/. [Accessed 16 June 2018].
• Presentation at ESSIR2017 on work by Radinsky, K., Svore, K.M., Dumais, S.T., Shokouhi, M.,
Teevan, J., Bocharov, A. and Horvitz, E., 2013. Behavioral dynamics on the web: Learning,
modeling, and prediction. ACM Transactions on Information Systems (TOIS), 31(3), p.16.
184. References
• The Stanford Question Answering Dataset. 2018. The Stanford
Question Answering Dataset. [ONLINE] Available
at: https://rajpurkar.github.io/SQuAD-explorer/.
• Trippas, J.R., Spina, D., Cavedon, L., Joho, H. and Sanderson, M., 2018.
Informing the Design of Spoken Conversational Search.
• https://medium.com/@ashishgupta031/sequence-aware-
reinforcement-learning-over-knowledge-graphs-a8af155e716c
• Jansen, B.J., Booth, D.L. and Spink, A., 2008. Determining the
informational, navigational, and transactional intent of Web
queries. Information Processing & Management, 44(3), pp.1251-
1266.
185. References
Radinsky, K., Svore, K.M., Dumais, S.T., Shokouhi, M., Teevan, J., Bocharov, A. and Horvitz, E., 2013.
Behavioral dynamics on the web: Learning, modeling, and prediction. ACM Transactions on Information
Systems (TOIS), 31(3), p.16
Sadikov, E., Madhavan, J. and Halevy, A., Google LLC, 2013. Clustering query
refinements by inferred user intent. U.S. Patent 8,423,538.
Official Google Webmaster Central Blog. 2019. Official Google Webmaster Central
Blog: Rolling out mobile-first indexing . [ONLINE] Available
at: https://webmasters.googleblog.com/2018/03/rolling-out-mobile-first-
indexing.html. [Accessed 25 September 2019].
Zhou, S., Cheng, K. and Men, L., 2017, April. The survey of large-scale query
classification. In AIP Conference Proceedings (Vol. 1834, No. 1, p. 040045). AIP
Publishing.
186. References
Search Engine Land. 2019. Starting July 1, all new sites will be indexed using Google's
mobile-first indexing - Search Engine Land. [ONLINE] Available
at: https://searchengineland.com/july-1-new-sites-will-be-indexed-using-googles-mobile-
first-indexing-317490. [Accessed 25 September 2019].
Teevan, J., Dumais, S.T. and Horvitz, E., 2005, August. Personalizing search via
automated analysis of interests and activities. In Proceedings of the 28th annual
international ACM SIGIR conference on Research and development in information
retrieval (pp. 449-456). ACM.
Nguyen, T., Rosenberg, M., Song, X., Gao, J., Tiwary, S., Majumder, R. and Deng, L.,
2016. MS MARCO: A Human-Generated MAchine Reading COmprehension Dataset.
198. MS MARCO Paper
• Nguyen, T., Rosenberg, M.,
Song, X., Gao, J., Tiwary, S.,
Majumder, R. and Deng, L.,
2016. MS MARCO: A Human-
Generated MAchine Reading
COmprehension Dataset.
202. There are
also several
types of
queries too
(Krisztian
Balog, ECIR,
2019)
Keyword queries (Normal keyword queries)
Keyword++ queries (Faceted / filtered
queries)
Zero-Query queries (User is the query)
Natural language queries
Structured queries (e.g. SQL)
203. 80% of all
queries are
information
al in nature
(Jansen et
al, 2008)
80%
10%
10%
Query Intent Split
Informational Transactional Navigational
213. What did you
really mean
when you
searched for
‘Easter’?
• Radinsky, K., Svore, K.M., Dumais, S.T., Shokouhi, M., Teevan, J.,
Bocharov, A. and Horvitz, E., 2013. Behavioral dynamics on the web:
Learning, modeling, and prediction. ACM Transactions on Information
Systems (TOIS), 31(3), p.16.
When did
you search
for ‘Easter’?
A few
weeks
before
Easter
A few days
before
Easter
During
Easter
What you mostly
meant
When is
Easter?
Things to do at
Easter
What is the
meaning of
Easter?
214. The problem is consistent
high precision is nowhere in
sight
215. And if we are to move into multi-device
ubiquitious search then…