In addition to natural language processing, other areas of artificial intelligence I have studied include
genetic algorithms and optimization [1, 2, 3, 4], game search [5, 6], computational economics and
finance [7, 8, 9, 10, 11, 12, 13], computational biology [14, 15, 16, 17, 18, 19], multimedia
processing [20, 21], machine learning algorithms and frameworks [22, 23, 24, 25, 26, 27].
My natural language research has spanned the three eras of rule-based, statistical, and neural systems. I started with Boris Katz’s rule-based natural language question answering system START (the longest running NLP system on the Internet) and developed its OmniBase component which allowed access to structured websites like IMDB and World Factbook using a uniform interface [28, 29, 30]. I later co-founded a company, Inquira Inc.,
which commercialized question answering technology for customer self-service applications of large
companies like Apple and Ebay [31, 32].
Natural languages are suffused with ambiguities and exceptions which makes development of robust rule-based systems difficult. Statistical models gradually replaced rule-based systems in the 1990s and 2000s as a more robust alternative. During this period I developed statistical models for supervised and unsupervised dependency parsing [33, 34], word sense
disambiguation and induction [35, 36, 37, 38, 39], child word category acquisition [40, 41, 42, 43],
morphological disambiguation [44, 45, 46], semantic role labeling [47], statistical language
modeling [48, 49], and machine translation [50, 51, 52, 53].
A major portion of this work
focused on unsupervised models because the large amounts of labeled data required for
supervised models are expensive to collect, difficult to get agreement on, and not required by
infants learning language. Nevertheless, supervised models play an important role in today’s
NLP applications, so to help create labeled datasets and perform evaluations that push
the state-of-the-art forward, I co-organized the CoNLL-2007 Shared Task on Dependency
Parsing [54], the SemEval-2007 Shared Task on Classification of Semantic Relations between
Nominals [55, 56], the SemEval-2010 Shared Task on Parser Evaluation Using Textual
Entailments [57, 58], the SemEval-2012 and SemEval-2013 Semantic Evaluation Exercises
[59, 60].
In 2010s, neural network based natural language processing systems started catching up in
performance with their statistical counterparts. More importantly, layers of (morphological,
syntactic, semantic) representations designed by linguists and used to train statistical models
have been gradually replaced with features automatically learned by deep models from
data (A similar transition took place in computer vision where hand-designed HOG/SIFT type features have beenreplaced with convolutional layers trained from data).
Feature engineering no longer plays the central role it did with statistical models: discrete features are
replaced by continuous embedding vectors, hand-designed feature combinations or kernels are replaced
by adaptive basis functions automatically learned by neural networks. For the first time, it became
possible to train end-to-end models, e.g. neural machine translation or image captioning systems are
trained on nothing but example input-output pairs.
During this period, I developed a novel continuous representation of word context based on the
distribution of possible substitutes for a word rather than its neighbors. My students and I showed
that this “paradigmatic” word context representation generalized better and improved the
state-of-the-art on problems such as unsupervised part of speech induction [61, 62, 63], unsupervised
word sense induction [64] and semantic word similarity [65]. Using neural models with little feature
engineering we also developed a named entity recognizer that achieves state-of-the-art results in 7
languages [66] and a dependency parser that parsed 81 treebanks in 49 languages and was ranked 7th
out of the 33 systems participating in the CoNLL-2017 UD Shared Task [67]. Neural models generally
require more data compared to statistical models which poses a problem in low-resource
settings. We showed one way to mitigate this problem using transfer learning: a low-resource
Turkish-English machine translation model performs 50% better when initialized with weights from a
high-resource French-English model compared to random initialization [68]. Another disadvantage
of neural models is their lack of interpretability, which we tried to address in [69] where
we discovered hidden units that count various features in a sequence-to-sequence RNN
model.
I am most excited about the potential of deep neural network models for grounded
language learning, i.e. learning the meanings of words, phrases, and sentences by observing
natural interactions. In a preliminary study, we showed that a neural model can learn to
follow instructions for arranging blocks on a table-top by observing humans giving and following instructions [70]. I am currently running two funded projects to explore this
topic further [71, 72] and our ongoing studies are promising [73]. I suspect robust natural
language understanding systems of the future will be end-to-end trained rather than hand
engineered.
[1] Deniz Yuret and Michael de la Maza. Dynamic hill climbing: Overcoming the
limitations of optimization techniques. In The Second Turkish Symposium on Artificial
Intelligence and Neural Networks, 1993.
[2] Michael de la Maza and Deniz Yuret. Dynamic hill climbing. AI Expert, 1994.
[3] Deniz Yuret. From genetic algorithms to efficient optimization. Technical Report 1569,
MIT AI Laboratory, 1994.
[4] Michael de la Maza and Deniz Yuret. Seeing clearly: Medical imaging now and
tomorrow. In Clifford A. Pickover, editor, Future Health: Computers and Medicine in the
21st Century. St. Martin’s Press, 1995.
[5] Deniz Yuret. The principle of pressure in chess. In The Third Turkish Symposium on
Artificial Intelligence and Neural Networks (TAINN ’94), 1994.
[6] David Allen McAllester and Deniz Yuret. Alpha-beta-conspiracy search. ICGA
Journal, 25(1):16–35, 2002.
[7] Michael de la Maza and Deniz Yuret. A futures market simulation with non-rational
participants. In Rodney Allen Brooks and Pattie Maes, editors, Proceedings of the Fourth
International Workshop on the Synthesis and Simulation of Living Systems, 1994.
[8] Deniz Yuret and Michael de la Maza. A genetic algorithm system for predicting the
OEX. Technical Analysis of Stocks and Commodities, 1994.
[9] Michael de la Maza and Deniz Yuret. Experimenting with a market simulation. The
Magazine of Artificial Intelligence in Finance, 1(3), 1994.
[10] Michael de la Maza and Deniz Yuret. A model of stock market participants. In Jörg
Biethahn and Volker Nissen, editors, Evolutionary Algorithms in Management Applications.
Springer, 1995.
[11] Michael de la Maza and Deniz Yuret. Neural network applications: A critique. The
Magazine of Artificial Intelligence in Finance, 2(1), 1995.
[12] Michael de la Maza, Ayla Oğuş, and Deniz Yuret. How do firms transition between
monopoly and competitive behavior? An agent-based economic model. In Proceedings of
the Sixth International Conference on Artificial Life, 1998.
[13] Ayla Oğuş, Michael de la Maza, and Deniz Yuret. Modeling the economics of internet
companies. In Computing in Economics and Finance, Proceedings of the Fifth International
Conference of the Society for Computational Economics, 1999.
[14] Özlem Keskin, Deniz Yuret, Attila Gürsoy, Metin Türkay, and Burak Erman.
Relationships between amino acid sequence and backbone torsion angle preferences.
Proteins: Structure, Function, and Bioinformatics, 55(4):992–998, June 2004.
[15] Ersin Yurtsever, Deniz Yuret, and Burak Erman. Quantum mechanical calculations
of tryptophan and comparison with conformations in native proteins. J. Phys. Chem. A,
110(51):13933–13938, December 2006.
[16] Alkan Kabakçıoğlu, Deniz Yuret, Mert Gür, and Burak Erman. Anharmonicity,
mode-coupling and entropy in a fluctuating native protein. Physical Biology, 7:046005,
October 2010.
[17] Onur Varol, Deniz Yuret, Burak Erman, and Alkan Kabakçıoğlu. Mode coupling
points to functionally important residues in myosin II. Proteins: Structure, Function, and
Bioinformatics, 82(9):1777–1786, September 2014.
[18] Alkan Kabakcioglu, Onur Varol, Deniz Yuret, and Burak Erman. Functionally
important residues from mode coupling during short-time protein dynamics. In APS
Meeting Abstracts, volume 1, page 48009, 2015.
[19] Onur Varol, Deniz Yuret, Burak Erman, and Alkan Kabakcioglu. Functionally
important residues from mode coupling during short-time protein dynamics. Biophysical
Journal, 108(2):377a, 2015.
[20] Barış Bozkurt, Ozan Baysal, and Deniz Yuret. A dataset and baseline system for
singing voice assessment. In The 13th International Symposium on Computer Music
Multidisciplinary Research (CMMR), September 2017.
[21] Saman Zia, Yücel Yemez, and Deniz Yuret. RGB-D object recognition using deep
convolutional neural networks. In The IEEE International Conference on Computer Vision
(ICCV), pages 896–903, October 2017.
[22] Deniz Yuret and Michael de la Maza. The greedy prepend algorithm for decision
list induction. In A. Levi et al., editors, ISCIS 2006, LNCS 4263, pages 37–46, Berlin
Heidelberg, November 2006. Springer-Verlag.
[23] Ergun Biçici and Deniz Yuret. Locally scaled density based clustering. In
B. Beliczynski et al., editors, ICANNGA 2007, Part I, LNCS 4431, pages 739–748, Berlin
Heidelberg, April 2007. Springer-Verlag.
[24] Deniz Yuret. Knet: beginning deep learning with 100 lines of Julia. In Machine Learning
Systems Workshop at NIPS 2016, December 2016.
[25] Enis Berk Çoban, Deniz Yuret, and Didem Unat. Multidimensional broadcast
operation on the GPU. In 5. Ulusal Yüksek Başarımlı Hesaplama Konferansı, İstanbul,
September 2017.
[26] Doğa Dikbayır, Enis Berk Çoban, İlker Kesen, Deniz Yuret, and Didem Unat.
Fast multidimensional reduction and broadcast operations on GPU for machine learning.
Concurrency and Computation: Practice and Experience, May 2018.
[27] Mike Innes, Deniz Yuret, et al. On machine learning and programming languages. In
SysML Conference, Stanford, CA, Feb 2018.
[28] Boris Katz, Deniz Yuret, et al. Blitz: a preprocessor for detecting context-independent
linguistic structures. In Proceedings of the 5th Pacific Rim International Conference on
Artificial Intelligence (PRICAI ’98), 1998.
[29] Boris Katz, Deniz Yuret, et al. Integrating web resources and lexicons into a natural
language query system. In Proceedings of the 6th IEEE International Conference on
Multimedia Computing and Systems (IEEE ICMCS’99), 1999.
[30] Boris Katz, Sue Felshin, Deniz Yuret, et al. Omnibase: Uniform access to heterogeneous
data for question answering. In NLDB 2002, LNCS 2553, pages 230–234. Springer-Verlag,
2002.
[31] Deniz Yuret. Method of utilizing implicit references to answer a query. US Patent
Number 6957213, Oct 2005.
[32] Edwin Riley Cooper, Gann Bierner, Laurel Kathleen Graham, Deniz Yuret,
James Charles Williams, and Filippo Beghelli. Ontology for use with a system, method,
and computer readable medium for retrieving information and response to a query. US
Patent Number 8612208, 9747390, Dec 2013.
[33] Deniz Yuret. Discovery of linguistic relations using lexical attraction. PhD thesis, MIT,
1998.
[34] Deniz Yuret. Dependency parsing as a classification problem. In Proceedings of the
Tenth Conference on Computational Natural Language Learning (CoNLL-X), June 2006.
[35] Özlem Uzuner, Boris Katz, and Deniz Yuret. Word sense disambiguation for
information retrieval. In Proceedings of the 1999 16th National Conference on Artificial
Intelligence (AAAI-99), 1999.
[36] Deniz Yuret. Some experiments with a Naive Bayes WSD system. In Rada Mihalcea
and Phil Edmonds, editors, Senseval-3: Third International Workshop on the Evaluation
of Systems for the Semantic Analysis of Text, pages 265–268, Barcelona, Spain, July 2004.
Association for Computational Linguistics.
[37] Ergun Biçici and Deniz Yuret. Clustering word pairs to answer analogy questions.
In Proceedings of the Fifteenth Turkish Symposium on Artificial Intelligence and Neural
Networks (TAINN 2006), June 2006.
[38] Deniz Yuret. KU: Word sense disambiguation by substitution. In Proceedings of the
Fourth International Workshop on Semantic Evaluations (SemEval-2007), pages 207–214,
Prague, Czech Republic, June 2007. Association for Computational Linguistics.
[39] Deniz Yuret and Mehmet Ali Yatbaz. The noisy channel model for unsupervised word
sense disambiguation. Computational Linguistics, 36(1):111–127, March 2010.
[40] Deniz Yuret, A. Engin Ural, F. Nihan Ketrez, Dilara Kocbas, and Aylin C. Kuntay.
Morphological cues vs. number of nominals in learning verb types from child-directed
speech. In Boston University Conference on Language Development (BUCLD33), October
2008.
[41] A. Engin Ural, Deniz Yuret, Nihan Ketrez, Dilara Kocbas, and Aylin Kuntay.
Morphological cues vs. number of nominals in learning verb types in turkish: Syntactic
bootstrapping mechanism revisited. Language and Cognitive Processes, 24(10):1393–1405,
December 2009.
[42] Mehmet Ali Yatbaz, Volkan Cirik, Aylin Küntay, and Deniz Yuret. Paradigmatic
representations outperform syntagmatic representations in distributional learning of
grammatical categories. In BUCLD, November 2014.
[43] Mehmet Ali Yatbaz, Volkan Cirik, Aylin Küntay, and Deniz Yuret. Learning
grammatical categories using paradigmatic representation: Substitute words for language
acquisition. In COLING, December 2016.
[44] Deniz Yuret and Ferhan Türe. Learning morphological disambiguation rules for
turkish. In HLT-NAACL 06, June 2006.
[45] Mehmet Ali Yatbaz and Deniz Yuret. Unsupervised morphological disambiguation
using statistical language models. In NIPS 2009 Workshop on Grammar Induction,
Representation of Language and Language Learning, Vancouver, Canada, December 2009.
[46] Deniz Yuret and Ergun Biçici. Modeling morphologically rich languages using split
words and unstructured dependencies. In ACL-IJCNLP, Singapore, August 2009.
[47] Deniz Yuret, Mehmet Ali Yatbaz, and Ahmet Engin Ural. Discriminative vs.
generative approaches in semantic role labeling. In Conference on Computational Natural
Language Learning (CoNLL), Manchaster, UK, Aug 2008.
[48] Deniz Yuret. Smoothing a tera-word language model. In Proceedings of ACL-08: HLT,
Short Papers, pages 141–144, Columbus, Ohio, June 2008. Association for Computational
Linguistics.
[49] Deniz Yuret. Fastsubs: An efficient and exact procedure for finding the most likely
lexical substitutes based on an n-gram language model. Signal Processing Letters, IEEE,
19(11):725–728, November 2012.
[50] E. Bicici and D. Yuret. L 1 regularized regression for reranking and system
combination in machine translation. In Proceedings of the Joint Fifth Workshop
on Statistical Machine Translation and MetricsMATR, pages 282–289. Association for
Computational Linguistics, July 2010.
[51] Ergun Biçici and Deniz Yuret. Instance selection for machine translation using feature
decay algorithms. In Proceedings of the Sixth Workshop on Statistical Machine Translation,
pages 272–283, Edinburgh, Scotland, July 2011. Association for Computational Linguistics.
[52] Ergun Biçici and Deniz Yuret. Regmt system for machine translation, system
combination, and evaluation. In Proceedings of the Sixth Workshop on Statistical Machine
Translation, pages 323–329, Edinburgh, Scotland, July 2011. Association for Computational
Linguistics.
[53] Ergun Biçici and Deniz Yuret. Optimizing instance selection for statistical machine
translation with feature decay algorithms. IEEE Transactions on Audio, Speech and
Language Processing, 23(2):339–350, February 2015.
[54] Joakim Nivre, Johan Hall, Sandra Kübler, Ryan McDonald, Jens Nilsson, Sebastian
Riedel, and Deniz Yuret, editors. The CoNLL 2007 Shared Task on Dependency Parsing,
Prague, Czech Republic, June 2007.
[55] Roxana Girju, Preslav Nakov, Vivi Nastase, Stan Szpakowicz, Peter Turney, and Deniz
Yuret. Semeval-2007 task 04: Classification of semantic relations between nominals. In
SemEval-2007: 4th International Workshop on Semantic Evaluations, June 2007.
[56] Roxana Girju, Preslav Nakov, Vivi Nastase, Stan Szpakowicz, Peter Turney, and Deniz
Yuret. Classification of semantic relations between nominals. Language Resources and
Evaluation, 43(2):105–121, June 2009.
[57] D. Yuret, A. Han, and Z. Turgut. Semeval-2010 task 12: Parser evaluation using
textual entailments. In Proceedings of the 5th International Workshop on Semantic
Evaluation, pages 51–56. Association for Computational Linguistics, July 2010.
[58] Deniz Yuret, Laura Rimell, and Aydin Han. Parser evaluation using textual entailments.
Language Resources and Evaluation, 47(3):639–659, September 2012.
[59] Deniz Yuret and Suresh Manandhar, editors. Proceedings of the Sixth International
Workshop on Semantic Evaluation (SemEval 2012), 2012.
[60] Deniz Yuret and Suresh Manandhar, editors. Proceedings of the Seventh International
Workshop on Semantic Evaluation (SemEval 2013), 2013.
[61] Mehmet Ali Yatbaz and Deniz Yuret. Unsupervised part of speech tagging using
unambiguous substitutes from a statistical language model. In Proceedings of the
23rd International Conference on Computational Linguistics: Posters, pages 1391–1398.
Association for Computational Linguistics, August 2010.
[62] Mehmet Ali Yatbaz, Enis Sert, and Deniz Yuret. Learning syntactic categories using
paradigmatic representations of word context. In Proceedings of the 2012 Conference on
Empirical Methods in Natural Language Processing (EMNLP-CONLL 2012), Jeju, Korea,
July 2012. Association for Computational Linguistics.
[63] Deniz Yuret, Mehmet Ali Yatbaz, and Enis Sert. Unsupervised instance-based part
of speech induction using probable substitutes. In Proceedings of COLING 2014, the
25th International Conference on Computational Linguistics: Technical Papers, pages
2303–2313, Dublin, Ireland, August 2014. Dublin City University and Association for
Computational Linguistics.
[64] Osman Başkaya, Enis Sert, Volkan Cirik, and Deniz Yuret. AI-KU: Using substitute
vectors and co-occurrence modeling for word sense induction and disambiguation. In Second
Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2: Proceedings
of the Seventh International Workshop on Semantic Evaluation (SemEval 2013), pages
300–306, Atlanta, Georgia, USA, June 2013. Association for Computational Linguistics.
[65] Oren Melamud, Ido Dagan, Jacob Goldberger, Idan Szpektor, and Deniz Yuret.
Probabilistic modeling of joint-context in distributional similarity. In Proceedings of the
Eighteenth Conference on Computational Natural Language Learning, pages 181–190, Ann
Arbor, Michigan, June 2014. Association for Computational Linguistics.
[66] Onur Kuru, Ozan Arkan Can, and Deniz Yuret. Charner: Character-level named entity
recognition. In COLING, December 2016.
[67] Ömer Kırnap, Berkay Furkan Önder, and Deniz Yuret. Parsing with context
embeddings. In Proceedings of the CoNLL 2017 Shared Task: Multilingual Parsing from
Raw Text to Universal Dependencies, pages 80–87, Vancouver, Canada, August 2017.
Association for Computational Linguistics.
[68] Barret Zoph, Deniz Yuret, Jon May, and Kevin Knight. Transfer learning for
low-resource neural machine translation. In Proceedings of the 2016 Conference on
Empirical Methods in Natural Language Processing, pages 1568–1575, Austin, Texas,
November 2016. Association for Computational Linguistics.
[69] Xing Shi, Kevin Knight, and Deniz Yuret. Why neural translations are the right
length. In Proceedings of the 2016 Conference on Empirical Methods in Natural
Language Processing, pages 2278–2282, Austin, Texas, November 2016. Association for
Computational Linguistics.
[70] Yonatan Bisk, Deniz Yuret, and Daniel Marcu. Natural language communication with
robots. In Proceedings of the 2016 Conference of the North American Chapter of the
Association for Computational Linguistics: Human Language Technologies, pages 751–761,
San Diego, California, June 2016. Association for Computational Linguistics.
[71] Yücel Yemez and Deniz Yuret. Dilbilimsel ve görsel i̇puçlarını birlikte kullanarak
gezinim dilinin öğrenilmesi. TÜBİTAK 1001 Project, 2015–2018.
[72] Luc De Raedt, Deniz Yuret, and Alessandro Saffiotti. Relational symbol grounding
through affordance learning (ReGROUND). CHIST-ERA Project on Human Language
Understanding: Grounding Language Learning, 2015–2018.
[73] Laura Antanas, Ozan Arkan Can, Jesse Davis, Luc De Raedt, Amy Loutfy, Andreas
Persson, Alessandro Saffiotti, Emre Ünal, Deniz Yuret, and Pedro Zuidberg dos Martires.
Relational symbol grounding through affordance learning: An overview of the ReGround
project. In Grounding Language Understanding (GLU 2017) ISCA Satellite Workshop of
Interspeech 2017. Stockholm University, August 2017.