<?xml version="1.0" encoding="utf-8"?>
<raweb xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="en" year="2018">
  <identification id="magnet" isproject="true">
    <shortname>MAGNET</shortname>
    <projectName>Machine Learning in Information Networks</projectName>
    <theme-de-recherche>Data and Knowledge Representation and Processing</theme-de-recherche>
    <domaine-de-recherche>Perception, Cognition and Interaction</domaine-de-recherche>
    <urlTeam>http://team.inria.fr/magnet/</urlTeam>
    <structure_exterieure type="Labs">
      <libelle>Centre de Recherche en Informatique, Signal et Automatique de Lille</libelle>
    </structure_exterieure>
    <structure_exterieure type="Organism">
      <libelle>CNRS</libelle>
    </structure_exterieure>
    <structure_exterieure type="Organism">
      <libelle>Université Charles de Gaulle (Lille 3)</libelle>
    </structure_exterieure>
    <header_dates_team>Creation of the Team: 2013 January 01, updated into Project-Team: 2016 May 01</header_dates_team>
    <LeTypeProjet>Project-Team</LeTypeProjet>
    <keywordsSdN>
      <term>A3.1. - Data</term>
      <term>A3.1.3. - Distributed data</term>
      <term>A3.1.4. - Uncertain data</term>
      <term>A3.4. - Machine learning and statistics</term>
      <term>A3.4.1. - Supervised learning</term>
      <term>A3.4.2. - Unsupervised learning</term>
      <term>A3.4.4. - Optimization and learning</term>
      <term>A3.5. - Social networks</term>
      <term>A3.5.1. - Analysis of large graphs</term>
      <term>A3.5.2. - Recommendation systems</term>
      <term>A4.8. - Privacy-enhancing technologies</term>
      <term>A9.4. - Natural language processing</term>
    </keywordsSdN>
    <keywordsSecteurs>
      <term>B1. - Life sciences</term>
      <term>B1.1.10. - Systems and synthetic biology</term>
      <term>B2. - Health</term>
      <term>B2.2.4. - Infectious diseases, Virology</term>
      <term>B2.3. - Epidemiology</term>
      <term>B2.4.1. - Pharmaco kinetics and dynamics</term>
      <term>B2.4.2. - Drug resistance</term>
      <term>B5.10. - Biotechnology</term>
      <term>B6.3. - Network functions</term>
      <term>B7.1.2. - Road traffic</term>
      <term>B8.3. - Urbanism and urban planning</term>
      <term>B9.5.1. - Computer science</term>
      <term>B9.5.4. - Chemistry</term>
      <term>B9.5.6. - Data science</term>
      <term>B9.6.8. - Linguistics</term>
      <term>B9.6.10. - Digital humanities</term>
      <term>B9.10. - Privacy</term>
    </keywordsSecteurs>
    <UR name="Lille"/>
  </identification>
  <team id="uid1">
    <person key="magnet-2018-idp154416">
      <firstname>Aurelien</firstname>
      <lastname>Bellet</lastname>
      <categoryPro>Chercheur</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, Researcher</moreinfo>
    </person>
    <person key="magnet-2018-idp156880">
      <firstname>Pascal</firstname>
      <lastname>Denis</lastname>
      <categoryPro>Chercheur</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, Researcher</moreinfo>
    </person>
    <person key="magnet-2018-idp159344">
      <firstname>Claudio</firstname>
      <lastname>Gentile</lastname>
      <categoryPro>Chercheur</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, Senior Researcher, from Oct 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp161824">
      <firstname>Jan</firstname>
      <lastname>Ramon</lastname>
      <categoryPro>Chercheur</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, Senior Researcher</moreinfo>
    </person>
    <person key="magnet-2018-idp164272">
      <firstname>Bert</firstname>
      <lastname>Cappelle</lastname>
      <categoryPro>Enseignant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille, Associate Professor, until Aug 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp166768">
      <firstname>Mikaela</firstname>
      <lastname>Keller</lastname>
      <categoryPro>Enseignant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille, Associate Professor</moreinfo>
    </person>
    <person key="magnet-2018-idp169248">
      <firstname>Marc</firstname>
      <lastname>Tommasi</lastname>
      <categoryPro>Enseignant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille, Team leader, Professor</moreinfo>
      <hdr>oui</hdr>
    </person>
    <person key="magnet-2018-idp172160">
      <firstname>Fabien</firstname>
      <lastname>Torre</lastname>
      <categoryPro>Enseignant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille, Associate Professor, until Sep 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp174656">
      <firstname>Fabio</firstname>
      <lastname>Vitale</lastname>
      <categoryPro>Enseignant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille, Associate Professor</moreinfo>
    </person>
    <person key="magnet-2018-idp177136">
      <firstname>Melissa</firstname>
      <lastname>Ailem</lastname>
      <categoryPro>PostDoc</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, until Sep 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp179600">
      <firstname>Thanh</firstname>
      <lastname>Le Van</lastname>
      <categoryPro>PostDoc</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, until Jun 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp182064">
      <firstname>Bo</firstname>
      <lastname>Li</lastname>
      <categoryPro>PostDoc</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille</moreinfo>
    </person>
    <person key="magnet-2018-idp184512">
      <firstname>Mahsa</firstname>
      <lastname>Asadi</lastname>
      <categoryPro>PhD</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, from Oct 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp186944">
      <firstname>Mathieu</firstname>
      <lastname>Dehouck</lastname>
      <categoryPro>PhD</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ des sciences et technologies de Lille</moreinfo>
    </person>
    <person key="magnet-2018-idp189408">
      <firstname>Onkar</firstname>
      <lastname>Pandit</lastname>
      <categoryPro>PhD</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria</moreinfo>
    </person>
    <person key="magnet-2018-idp191840">
      <firstname>Arijus</firstname>
      <lastname>Pleska</lastname>
      <categoryPro>PhD</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, from Oct 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp194272">
      <firstname>Brij Mohan Lal</firstname>
      <lastname>Srivastava</lastname>
      <categoryPro>PhD</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, from Oct 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp196704">
      <firstname>Mariana</firstname>
      <lastname>Vargas Vieyra</lastname>
      <categoryPro>PhD</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria</moreinfo>
    </person>
    <person key="magnet-2018-idp199136">
      <firstname>William</firstname>
      <lastname>de Vazelhes</lastname>
      <categoryPro>Technique</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria</moreinfo>
    </person>
    <person key="magnet-2018-idp191840">
      <firstname>Arijus</firstname>
      <lastname>Pleska</lastname>
      <categoryPro>Technique</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, until Sep 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp204064">
      <firstname>César</firstname>
      <lastname>Sabater</lastname>
      <categoryPro>Technique</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria</moreinfo>
    </person>
    <person key="magnet-2018-idp206528">
      <firstname>Carlos</firstname>
      <lastname>Zubiaga Pena</lastname>
      <categoryPro>Technique</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria</moreinfo>
    </person>
    <person key="magnet-2018-idp208992">
      <firstname>Igor</firstname>
      <lastname>Axinti</lastname>
      <categoryPro>Stagiaire</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ des sciences et technologies de Lille, from Mar 2018 until Aug 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp211520">
      <firstname>Antoine</firstname>
      <lastname>Capriski</lastname>
      <categoryPro>Stagiaire</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, from May 2018 until Aug 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp214000">
      <firstname>Arthur</firstname>
      <lastname>d'Azemar</lastname>
      <categoryPro>Stagiaire</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ des sciences et technologies de Lille, from Mar 2018 until Aug 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp216528">
      <firstname>Alexandre</firstname>
      <lastname>Huat</lastname>
      <categoryPro>Stagiaire</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria, from Mar 2018 until Aug 2018</moreinfo>
    </person>
    <person key="bonus-2018-idp154912">
      <firstname>Julie</firstname>
      <lastname>Jonas</lastname>
      <categoryPro>Assistant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Inria</moreinfo>
    </person>
    <person key="magnet-2018-idp221472">
      <firstname>Tejas</firstname>
      <lastname>Kulkarni</lastname>
      <categoryPro>Visiteur</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Warwick University, from May 2018 until Aug 2018</moreinfo>
    </person>
    <person key="magnet-2018-idp223968">
      <firstname>Remi</firstname>
      <lastname>Gilleron</lastname>
      <categoryPro>CollaborateurExterieur</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille</moreinfo>
      <hdr>oui</hdr>
    </person>
    <person key="orpailleur-2018-idp186432">
      <firstname>Joël</firstname>
      <lastname>Legrand</lastname>
      <categoryPro>Enseignant</categoryPro>
      <research-centre>Lille</research-centre>
      <moreinfo>Univ of Lille, Associate Professor</moreinfo>
    </person>
  </team>
  <presentation id="uid2">
    <bodyTitle>Overall Objectives</bodyTitle>
    <subsection id="uid3" level="1">
      <bodyTitle>Presentation</bodyTitle>
      <p><span class="smallcap" align="left">Magnet</span> is a research group that aims to design new machine learning
based methods geared towards mining information
networks. Information networks are large collections of
interconnected data and documents like citation networks and blog
networks among others. Our goal is to propose new prediction
methods for texts and networks of texts based on machine learning
algorithms in graphs. Such algorithms include node and link
classification, link prediction, clustering and probabilistic
modeling of graphs. We aim to tackle real-world problems such as
browsing, monitoring and recommender systems, and more broadly
information extraction in information networks. Application domains
cover natural language processing, social networks for cultural data
and e-commerce, and biomedical informatics.</p>
    </subsection>
  </presentation>
  <fondements id="uid4">
    <bodyTitle>Research Program</bodyTitle>
    <subsection id="uid5" level="1">
      <bodyTitle>Introduction</bodyTitle>
      <p>The main objective of <span class="smallcap" align="left">Magnet</span> is to develop original machine
learning methods for networked data in order to build applications
like browsing, monitoring and recommender systems, and more broadly
information extraction in information networks. We consider
information networks in which the data consist of both feature
vectors and texts. We model such networks as (multiple)
(hyper)graphs wherein nodes correspond to entities (documents, spans
of text, users, ...) and edges correspond to relations between
entities (similarity, answer, co-authoring, friendship, ...). Our
main research goal is to propose new on-line and batch learning
algorithms for various problems (node classification / clustering,
link classification / prediction) which exploit the relationships
between data entities and, overall, the graph topology. We are also
interested in searching for the best hidden graph structure to be
generated for solving a given learning task. Our research will be
based on generative models for graphs, on machine learning for
graphs and on machine learning for texts. The challenges are the
dimensionality of the input space, possibly the dimensionality of
the output space, the high level of dependencies between the data,
the inherent ambiguity of textual data and the limited amount of
human labeling. An additional challenge will be to design scalable
methods for large information networks. Hence, we will explore how
sampling, randomization and active learning can be leveraged to
improve the scalability of the proposed algorithms.</p>
      <p>Our research
program is organized according to the following questions:</p>
      <orderedlist>
        <li id="uid6">
          <p noindent="true">How to go beyond vectorial classification models in Natural
Language Processing (NLP) tasks?</p>
        </li>
        <li id="uid7">
          <p noindent="true">How to adaptively build graphs with respect to the given tasks?
How to create networks from observations of information diffusion
processes?</p>
        </li>
        <li id="uid8">
          <p noindent="true">How to design methods able to achieve a good trade-off between predictive
accuracy and computational complexity?</p>
        </li>
        <li id="uid9">
          <p noindent="true">How to go beyond strict node homophilic/similarity assumptions
in graph-based learning methods?</p>
        </li>
      </orderedlist>
    </subsection>
    <subsection id="uid10" level="1">
      <bodyTitle>Beyond Vectorial Models for NLP</bodyTitle>
      <p>One of our overall research objectives is to derive graph-based
machine learning algorithms for natural language and text information
extraction tasks. This section discusses the motivations behind the
use of graph-based ML approaches for these tasks, the main challenges
associated with it, as well as some concrete projects. Some of the
challenges go beyond NLP problems and will be further developed in the
next sections. An interesting aspect of the project is that we
anticipate some important cross-fertilizations between NLP and ML
graph-based techniques, with NLP not only benefiting from but also
pushing ML graph-based approaches into new directions.</p>
      <p>Motivations for resorting to graph-based algorithms for texts are at
least threefold. First, online texts are organized in networks. With
the advent of the web, and the development of forums, blogs, and
micro-blogging, and other forms of social media, text productions have
become strongly connected.
Interestingly, NLP research has been rather slow in coming to
terms with this situation, and most of the literature still focus on document-based
or sentence-based predictions (wherein inter-document or
inter-sentence structure is not exploited). Furthermore, several
multi-document tasks exist in NLP (such as multi-document
summarization and cross-document coreference resolution), but most
existing work typically ignore document boundaries and simply apply a
document-based approach, therefore failing to take advantage of the
multi-document dimension <ref xlink:href="#magnet-2018-bid0" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid1" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p>A second motivation comes from the fact that most (if not all) NLP
problems can be naturally conceived as graph problems. Thus, NLP tasks
often involve discovering a relational structure over a set of text
spans (words, phrases, clauses, sentences, etc.). Furthermore, the
<i>input</i> of numerous NLP tasks is also a graph; indeed, most
end-to-end NLP systems are conceived as pipelines wherein the output
of one processor is in the input of the next. For instance, several
tasks take POS tagged sequences or dependency trees as input. But this
structured input is often converted to a vectorial form, which
inevitably involves a loss of information.</p>
      <p>Finally, graph-based representations and learning methods
appear to address some core problems faced by NLP, such as the fact
that textual data are typically not independent and identically
distributed, they often live on a manifold, they involve very high
dimensionality, and their annotations is costly and scarce. As such,
graph-based methods represent an interesting alternative to, or at least
complement, structured prediction methods (such as CRFs or
structured SVMs) commonly used within NLP.
Graph-based methods, like label propagation,
have also been shown to be very effective in semi-supervised settings,
and have already given some positive results on a few NLP tasks
<ref xlink:href="#magnet-2018-bid2" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid3" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p>Given the above motivations, our first line of research will be to
investigate how one can leverage an underlying network structure
(e.g., hyperlinks, user links) between documents, or text spans in
general, to enhance prediction performance for several NLP tasks. We
think that a “network effect”, similar to the one that took place in
Information Retrieval (with the Page Rank algorithm), could also
positively impact NLP research. A few recent papers have already
opened the way, for instance in attempting to exploit Twitter follower
graph to improve sentiment classification  <ref xlink:href="#magnet-2018-bid4" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p>Part of the challenge here will be to investigate how
adequately and efficiently one can model these problems as instances
of more general graph-based problems, such as node
clustering/classification or link prediction discussed in the next
sections. In a few
cases, like text classification or sentiment analysis, graph modeling
appears to be straightforward: nodes correspond to texts (and
potentially users), and edges are given by relationships like
hyperlinks, co-authorship, friendship, or thread
membership. Unfortunately, modeling NLP problems as networks is not
always that obvious. From the one hand, the right level of
representation will probably vary depending on the task at hand: the
nodes will be sentences, phrases, words, etc. From the other hand, the
underlying graph will typically not be given a priori, which in turn
raises the question of how we construct it.
A preliminary discussion of the issue of optimal graph
construction for semi-supervised learning in NLP is given in
<ref xlink:href="#magnet-2018-bid2" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid5" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>. We identify the
issue of adaptive graph construction as an important scientific
challenge for machine learning on graphs in general, and we will
discuss it further in Section <ref xlink:href="#uid11" location="intern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p>As noted above, many NLP tasks have been recast as structured
prediction problems, allowing to capture (some of the) output
dependencies.
How to best combine structured output and graph-based ML
approaches is another challenge that we intend to address. We will
initially investigate this question within a semi-supervised context,
concentrating on graph regularization and graph propagation
methods. Within such approaches, labels are typically binary or in a
small finite set. Our objective is to explore how one
propagates an exponential number of <i>structured labels</i> (like a
sequence of tags or a dependency tree) through graphs. Recent attempts
at blending structured output models with graph-based models are
investigated in <ref xlink:href="#magnet-2018-bid3" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid6" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>. Another
related question that we will address in this context is how does one
learn with <i>partial labels</i> (like partially specified tag
sequence or tree) and use the graph structure to complete the output
structure. This last question is very relevant to NLP problems where
human annotations are costly; being able to learn from partial
annotations could therefore allow for more targeted annotations and in
turn reduced costs <ref xlink:href="#magnet-2018-bid7" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p>The NLP tasks we will mostly focus on are coreference resolution and
entity linking, temporal structure prediction, and discourse
parsing. These tasks will be envisioned in both document and
cross-document settings, although we expect to exploit inter-document
links either way. Choices for these particular tasks is guided by the
fact that they are still open problems for the NLP community, they
potentially have a high impact for industrial applications (like
information retrieval, question answering, etc.), and we already have
some expertise on these tasks in the team (see for instance <ref xlink:href="#magnet-2018-bid8" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid9" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid10" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>). As a midterm goal, we also
plan to work on tasks more directly relating to micro-blogging, such
sentiment analysis and the automatic thread structuring of technical
forums; the latter task is in fact an instance of rhetorical structure
prediction <ref xlink:href="#magnet-2018-bid11" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.
We have already initiated some work on the coreference resolution with graph-based learning, by casting the problem as an instance of spectral clustering <ref xlink:href="#magnet-2018-bid10" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
    </subsection>
    <subsection id="uid11" level="1">
      <bodyTitle>Adaptive Graph Construction</bodyTitle>
      <p>In most applications, edge weights are computed through a complex
data modeling process and convey crucially important information for
classifying nodes, making it possible to infer information
related to each data sample even exploiting the graph topology solely.
In fact, a widespread approach to several
classification problems is to represent the data through an undirected
weighted graph in which edge weights quantify the similarity between
data points. This technique for coding input data has been applied to
several domains, including classification of genomic data <ref xlink:href="#magnet-2018-bid12" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, face recognition <ref xlink:href="#magnet-2018-bid13" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, and text categorization <ref xlink:href="#magnet-2018-bid14" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p>In some cases, the full adjacency matrix is generated by employing
suitable similarity functions chosen through a deep understanding of
the problem structure. For example for the TF-IDF representation of documents,
the affinity between pairs of samples is often estimated through the
cosine measure or the <formula type="inline"><math xmlns="http://www.w3.org/1998/Math/MathML" overflow="scroll"><msup><mi>χ</mi><mn>2</mn></msup></math></formula> distance. After the generation of the
full adjacency matrix, the second phase for obtaining the final graph
consists in an edge sparsification/reweighting operation. Some of the
edges of the clique obtained in the first step are pruned and the
remaining ones can be reweighted to meet the specific requirements of
the given classification problem. Constructing a graph with these
methods obviously entails various kinds of loss of
information. However, in problems like node classification, the use of
graphs generated from several datasets can lead to an improvement in
accuracy ( <ref xlink:href="#magnet-2018-bid15" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid16" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid17" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
Hence, the transformation of a dataset into a graph may, at least in
some cases, partially remove various kinds of irregularities present
in the original datasets, while keeping some of the most useful
information for classifying the data samples. Moreover, it is often
possible to accomplish classification tasks on the obtained graph
using a running time remarkably lower than is needed by algorithms
exploiting the initial datasets, and a suitable sparse graph
representation can be seen as a compressed version of the original
data. This holds even when input data are provided in a online/stream
fashion, so that the resulting graph evolves over time.</p>
      <p>In this project we will address the problem of adaptive graph
construction towards several directions. The first one is about how to choose the best similarity measure given the objective learning
task. This question is related to the question of metric and similarity learning
( <ref xlink:href="#magnet-2018-bid18" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid19" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>) which has not been considered in the
context of graph-based learning. In the context of structured
prediction, we will develop approaches where output structures are
organized in graphs whose similarity is given by top-<formula type="inline"><math xmlns="http://www.w3.org/1998/Math/MathML" overflow="scroll"><mi>k</mi></math></formula> outcomes of
greedy algorithms.</p>
      <p>A different way we envision adaptive graph construction is in the
context of semi-supervised learning. Partial supervision can take
various forms and an interesting and original setting is governed by
two currently studied applications: detection of brain anomaly from
connectome data and polls recommendation in marketing. Indeed, for
these two applications, a partial knowledge of the information
diffusion process can be observed while the network is unknown or only
partially known. An objective is to construct (or complete) the
network structure from some local diffusion information. The problem
can be formalized as a graph construction problem from partially
observed diffusion processes. It has been studied very recently in
<ref xlink:href="#magnet-2018-bid20" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>. In our case, the originality comes
either from the existence of different sources of observations or from
the large impact of node contents in the network.</p>
      <p>We will study how to combine graphs defined by networked data and
graphs built from flat data to solve a given task. This is of major
importance for information networks because, as said above, we will
have to deal with multiple relations between entities (texts, spans of
texts, ...) and also use textual data and vectorial data.</p>
    </subsection>
    <subsection id="uid12" level="1">
      <bodyTitle>Prediction on Graphs and Scalability</bodyTitle>
      <p>As stated in the previous sections, graphs as complex objects provide
a rich representation of data. Often enough the data is only partially
available and the graph representation is very helpful in predicting
the unobserved elements. We are interested in problems where the
complete structure of the graph needs to be recovered and only a
fraction of the links is observed. The link prediction problem falls
into this category. We are also interested in the recommendation and
link classification problems which can be seen as graphs where the
structure is complete but some labels on the links (weights or signs)
are missing. Finally we are also interested in labeling the nodes of
the graph, with class or cluster memberships or with a real value,
provided that we have (some information about) the labels for some of
the nodes.</p>
      <p>The semi-supervised framework will be also considered. A midterm
research plan is to study how graph regularization models help
for structured prediction problems. This question will be studied in
the context of NLP tasks, as noted in
Section <ref xlink:href="#uid10" location="intern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, but we also plan to develop
original machine learning algorithms that have a more general
applicability. Inputs are networks whose nodes (texts) have to be
labeled by structures. We assume that structures lie in some manifold
and we want to study how labels can propagate in the network. One
approach is to find a smooth labeling function corresponding to an
harmonic function on both manifolds in input and output.</p>
      <p>Scalability is one of the main issues in the design of new prediction
algorithms working on networked data. It has gained more and more
importance in recent years, because of the growing size of the most
popular networked data that are now used by millions of people. In such contexts, learning algorithms whose computational complexity scales
quadratically, or slower, in the number of considered data objects
(usually nodes or edges, depending on the task) should be
considered impractical.</p>
      <p>These observations lead to the idea of using graph sparsification
techniques in order to work on a part of the original network for getting
results that can be easily extended and used for the whole original
input. A sparsified version of the original graph can often be seen as a
subset of the initial input, i.e. a suitably selected input subgraph which
forms the training set (or, more in general, it is included in the training
set). This holds even for the active setting.
A simple example could be to find a spanning tree of the input graph,
possibly using randomization techniques, with properties such that we
are allowed to obtain interesting results for the initial graph
dataset. We have started to explore this research direction for
instance in <ref xlink:href="#magnet-2018-bid21" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.</p>
      <p spacebefore="3.0pt">At the level of mathematical foundations, the key issue to be
addressed in the study of (large-scale) random networks also concerns
the segmentation of network data into sets of independent and
identically distributed observations. If we identify the data sample
with the whole network, as it has been done in previous approaches
<ref xlink:href="#magnet-2018-bid22" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, we typically end up with a set of
observations (such as nodes or edges) which are highly interdependent
and hence overly violate the classic i.i.d. assumption. In this case,
the data scale can be so large and the range of correlations can be so
wide, that the cost of taking into account the whole data and their
dependencies is typically prohibitive. On the contrary, if we focus
instead on a set of subgraphs independently drawn from a (virtually
infinite) target network, we come up with a set of independent and
identically distributed observations—namely the subgraphs
themselves, where subgraph sampling is the underlying ergodic process
<ref xlink:href="#magnet-2018-bid23" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>. Such an approach is one principled direction
for giving novel statistical foundations to random network
modeling. At the same time, because one shifts the focus from the
whole network to a set of subgraphs, complexity issues can be
restricted to the number of subgraphs and their size. The latter
quantities can be controlled much more easily than the overall network
size and dependence relationships, thus allowing to tackle scalability
challenges through a radically redesigned approach.</p>
      <p>Another way to tackle scalability problems is to exploit the inherent decentralized nature of very large graphs. Indeed, in many situations very large graphs are the abstract view of the digital activities of a very large set of users equipped with their own device. Nowadays, smartphones, tablets and even sensors have storage and computation power and gather a lot of data that serve to analytics, prediction, suggestion and personalized recommendation. Gathering all user data in large data centers is costly because it requires oversized infrastructures with huge energy consumption and large bandwidth networks. Even though cloud architectures can optimize such infrastructures, data concentration is also prone to security leaks, lost of privacy and data governance for end users.
The alternative we have started to develop in Magnet is to devise decentralized, private and personalized machine learning algorithms so that they can be deployed in the personal devices. The key challenges are therefore to learn in a collaborative way in a network of learners and to preserve privacy and control on personal data.</p>
    </subsection>
    <subsection id="uid13" level="1">
      <bodyTitle>Beyond Homophilic Relationships</bodyTitle>
      <p>In many cases, algorithms for solving node classification
problems are driven by the following assumption: linked entities tend
to be assigned to the same class. This assumption, in the context of
social networks, is known as homophily
( <ref xlink:href="#magnet-2018-bid24" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>, <ref xlink:href="#magnet-2018-bid25" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>) and involves ties of every
type, including friendship, work, marriage, age, gender, and so on. In
social networks, homophily naturally implies that a set of individuals
can be parted into subpopulations that are more cohesive. In fact, the
presence of homogeneous groups sharing common interests is a key reason for affinity among interconnected individuals,
which suggests that, in spite of its simplicity, this principle turns
out to be very powerful for node classification problems in general
networks.</p>
      <p>Recently, however, researchers have started to consider networked data
where connections may also carry a negative meaning. For instance,
disapproval or distrust in social networks, negative endorsements on the
Web.
Although the introduction of signs on graph edges appears like a small
change from standard weighted graphs, the resulting mathematical model,
called signed graphs, has an unexpectedly rich additional complexity. For
example, their spectral properties, which essentially all
sophisticated node classification algorithms rely on, are different and
less known than those of graphs. Signed graphs
naturally lead to a specific inference problem that we have discussed in
previous sections: link classification. This is the problem of predicting
signs of links in a given graph. In online social networks, this may be
viewed as a form of sentiment analysis, since we would like to semantically
categorize the relationships between individuals.</p>
      <p spacebefore="3.0pt">Another way to go
beyond homophily between entities will be studied using our recent
model of hypergraphs with bipartite hyperedges
<ref xlink:href="#magnet-2018-bid26" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>.
A bipartite hyperedge
connects two ends which are disjoint subsets of nodes. Bipartite
hyperedges is a way to relate two collections of (possibly
heterogeneous) entities represented by nodes. In the NLP setting,
while hyperedges can be used to model bags of words, bipartite
hyperedges are associated with relationships between bags of
words. But each end of bipartite hyperedges is also a way to represent
complex entities, gathering several attribute values (nodes) into
hyperedges viewed as records. Our hypergraph notion naturally extends
directed and undirected weighted graph. We have defined a spectral
theory for this new class of hypergraphs and opened a way to smooth
labeling on sets of nodes. The weighting scheme allows to weigh the
participation of each node to the relationship modeled by bipartite
hyperedges accordingly to an equilibrium condition. This condition provides a competition between nodes
in hyperedges and allows interesting modeling properties that go
beyond homophily and similarity over nodes (the theoretical analysis of
our hypergraphs exhibits tight relationships with signed
graphs). Following this competition idea, bipartite hyperedges are
like matches between two teams and examples of applications are team
creation. The basic tasks we are interested in are hyperedge
classification, hyperedge prediction, node weight prediction. Finally,
hypergraphs also represent a way to summarize or compress large graphs
in which there exists highly connected couples of (large) subsets of
nodes.</p>
    </subsection>
  </fondements>
  <domaine id="uid14">
    <bodyTitle>Application Domains</bodyTitle>
    <subsection id="uid15" level="1">
      <bodyTitle>Domain 1</bodyTitle>
      <p>Our main targeted applications are browsing, monitoring, recommending and mining in information networks. The learning tasks considered in the project such as node clustering, node and link classification and link prediction are likely to yield important improvements in these applications. Application
domains cover social networks for cultural data and e-commerce, and biomedical informatics.</p>
      <p>We also target applications related to decentralized learning and privacy preserving systems when users or devices are interconnected in large networks. We develop solutions based on urban and mobility data where privacy is a specific requirement.</p>
    </subsection>
  </domaine>
  <highlights id="uid16">
    <bodyTitle>Highlights of the Year</bodyTitle>
    <subsection id="uid17" level="1">
      <bodyTitle>Highlights of the Year</bodyTitle>
      <simplelist>
        <li id="uid18">
          <p noindent="true">Strengthening of the privacy aware machine learning activity
with a new associate team with the Alan Turing Institute and the
organization of a workshop at NeurIPS (formerly NIPS).</p>
        </li>
        <li id="uid19">
          <p noindent="true">New collaboration with Multispeech (Inria Nancy) on
decentralized and private machine learning for speech processing
leading to an ANR and an H2020 project.</p>
        </li>
      </simplelist>
      <subsection id="uid20" level="2">
        <bodyTitle>Awards</bodyTitle>
        <p><span class="smallcap" align="left">Aurélien Bellet</span> received a best reviewer award (top 200 out of 3000) at the
conference NeurIPS 2018.
<span class="smallcap" align="left">Pascal Denis</span> received a Distinguished Senior Program Committee award at IJCAI-ECAI 2018.</p>
      </subsection>
    </subsection>
  </highlights>
  <logiciels id="uid21">
    <bodyTitle>New Software and Platforms</bodyTitle>
    <subsection id="uid22" level="1">
      <bodyTitle>CoRTeX</bodyTitle>
      <p>
        <i>Python library for noun phrase COreference Resolution in natural language TEXts</i>
      </p>
      <p noindent="true"><span class="smallcap" align="left">Keyword:</span> Natural language processing</p>
      <p noindent="true"><span class="smallcap" align="left">Functional Description:</span> CoRTex is a LGPL-licensed Python library for Noun Phrase coreference resolution in natural language texts. This library contains implementations of various state-of-the-art coreference resolution algorithms, including those developed in our research. In addition, it provides a set of APIs and utilities for text pre-processing, reading the CONLL2012 and CONLLU annotation formats, and performing evaluation, notably based on the main evaluation metrics (MUC, B-CUBED, and CEAF). As such, CoRTex provides benchmarks for researchers working on coreference resolution, but it is also of interest for developers who want to integrate a coreference resolution within a larger platform. It currently supports use of the English or French language.</p>
      <simplelist>
        <li id="uid23">
          <p noindent="true">Participant: Pascal Denis</p>
        </li>
        <li id="uid24">
          <p noindent="true">Partner: Orange Labs</p>
        </li>
        <li id="uid25">
          <p noindent="true">Contact: Pascal Denis</p>
        </li>
        <li id="uid26">
          <p noindent="true">URL: <ref xlink:href="https://gitlab.inria.fr/magnet/CoRTeX" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>gitlab.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>magnet/<allowbreak/>CoRTeX</ref></p>
        </li>
      </simplelist>
    </subsection>
    <subsection id="uid27" level="1">
      <bodyTitle>Mangoes</bodyTitle>
      <p>
        <i>MAgnet liNGuistic wOrd vEctorS</i>
      </p>
      <p noindent="true"><span class="smallcap" align="left">Keywords:</span> Word embeddings - NLP</p>
      <p noindent="true"><span class="smallcap" align="left">Functional Description:</span> Process textual data and compute vocabularies and co-occurrence matrices. Input data should be raw text or annotated text.
Compute word embeddings with different state-of-the art unsupervised methods.
Propose statistical and intrinsic evaluation methods, as well as some visualization tools.</p>
      <simplelist>
        <li id="uid28">
          <p noindent="true">Contact: Nathalie Vauquier</p>
        </li>
        <li id="uid29">
          <p noindent="true">URL: <ref xlink:href="https://gitlab.inria.fr/magnet/mangoes" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>gitlab.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>magnet/<allowbreak/>mangoes</ref></p>
        </li>
      </simplelist>
    </subsection>
    <subsection id="uid30" level="1">
      <bodyTitle>metric-learn</bodyTitle>
      <p><span class="smallcap" align="left">Keywords:</span> Machine learning - Python - Metric learning</p>
      <p noindent="true"><span class="smallcap" align="left">Functional Description:</span> Distance metrics are widely used in the machine learning literature. Traditionally, practicioners would choose a standard distance metric (Euclidean, City-Block, Cosine, etc.) using a priori knowledge of the domain. Distance metric learning (or simply, metric learning) is the sub-field of machine learning dedicated to automatically constructing optimal distance metrics.</p>
      <p>This package contains efficient Python implementations of several popular metric learning algorithms.</p>
      <simplelist>
        <li id="uid31">
          <p noindent="true">Partner: Parietal</p>
        </li>
        <li id="uid32">
          <p noindent="true">Contact: William De Vazelhes</p>
        </li>
        <li id="uid33">
          <p noindent="true">URL: <ref xlink:href="https://github.com/metric-learn/metric-learn" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>github.<allowbreak/>com/<allowbreak/>metric-learn/<allowbreak/>metric-learn</ref></p>
        </li>
      </simplelist>
    </subsection>
    <subsection id="uid34" level="1">
      <bodyTitle>MyLocalInfo</bodyTitle>
      <p><span class="smallcap" align="left">Keywords:</span> Privacy - Machine learning - Statistics</p>
      <p noindent="true"><span class="smallcap" align="left">Functional Description:</span> Decentralized algorithms for machine learning and inference tasks which
(1) perform as much computation as possible locally and (2) ensure privacy and
security by avoiding personal data leaves devices.</p>
      <simplelist>
        <li id="uid35">
          <p noindent="true">Contact: Nathalie Vauquier</p>
        </li>
        <li id="uid36">
          <p noindent="true">URL: <ref xlink:href="https://gitlab.inria.fr/magnet/mylocalinfo" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>gitlab.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>magnet/<allowbreak/>mylocalinfo</ref></p>
        </li>
      </simplelist>
    </subsection>
  </logiciels>
  <resultats id="uid37">
    <bodyTitle>New Results</bodyTitle>
    <subsection id="uid38" level="1">
      <bodyTitle>On the Bernstein-Hoeffding Method</bodyTitle>
      <p>We consider extensions of Hoeffding's “exponential method”
approach for obtaining upper estimates on the probability that a sum
of independent and bounded random variables is significantly larger
than its mean. We show that the exponential function in Hoeffding's
approach can be replaced with any function which is non-negative,
increasing and convex. As a result we generalize and improve upon
Hoeffding's inequality. Our approach allows to obtain “missing
factors” in Hoeffding's inequality. The later result is a rather
weaker version of a theorem that is due to Michel
Talagrand. Moreover, we characterize the class of functions with
respect to which our method yields optimal concentration
bounds. Finally, using ideas from the theory of Bernstein
polynomials, we show that similar ideas apply under information on
higher moments of the random
variables (<ref xlink:href="#magnet-2018-bid27" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid39" level="1">
      <bodyTitle>IncGraph: Incremental graphlet counting for topology optimisation </bodyTitle>
      <p>Graphlets are small network patterns that can be counted in order to
characterise the structure of a network (topology). As part of a
topology optimisation process, one could use graphlet counts to
iteratively modify a network and keep track of the graphlet counts,
in order to achieve certain topological properties. Up until now,
however, graphlets were not suited as a metric for performing
topology optimisation; when millions of minor changes are made to
the network structure it becomes computationally intractable to
recalculate all the graphlet counts for each of the edge
modifications. We propose IncGraph, a
method for calculating the differences in graphlet counts with
respect to the network in its previous state, which is much more
efficient than calculating the graphlet occurrences from scratch at
every edge modification made. In comparison to static counting
approaches, our findings show IncGraph reduces the execution time by
several orders of magnitude. The usefulness of this approach was
demonstrated by developing a graphlet-based metric to optimise gene
regulatory networks. IncGraph is able to quickly quantify the
topological impact of small changes to a network, which opens novel
research opportunities to study changes in topologies in evolving or
online networks, or develop graphlet-based criteria for topology
optimisation. IncGraph is freely available as an open-source R
package on CRAN (incgraph). The development version is also
available on GitHub (rcannood/incgraph) (<ref xlink:href="#magnet-2018-bid28" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid40" level="1">
      <bodyTitle>Graph sampling with applications to estimating the number of pattern embeddings and the parameters of a statistical relational model </bodyTitle>
      <p>Counting the number of times a pattern occurs in a database is a
fundamental data mining problem. It is a subroutine in a diverse set
of tasks ranging from pattern mining to supervised learning and
probabilistic model learning. While a pattern and a database can
take many forms, this paper focuses on the case where both the
pattern and the database are graphs (networks). Unfortunately, in
general, the problem of counting graph occurrences is
#P-complete. In contrast to earlier work, which focused on exact
counting for simple (i.e., very short) patterns, we present a
sampling approach for estimating the statistics of larger graph
pattern occurrences. We perform an empirical evaluation on synthetic
and real-world data that validates the proposed algorithm,
illustrates its practical behavior and provides insight into the
trade-off between its accuracy of estimation and computational
efficiency (<ref xlink:href="#magnet-2018-bid29" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid41" level="1">
      <bodyTitle>A machine learning based framework to
identify and classify long terminal repeat retrotransposons</bodyTitle>
      <p>Transposable elements (TEs) are repetitive nucleotide sequences that
make up a large portion of eukaryotic genomes. They can move and
duplicate within a genome, increasing genome size and contributing
to genetic diversity within and across species. Accurate
identification and classification of TEs present in a genome is an
important step towards understanding their effects on genes and
their role in genome evolution. We introduce TE-LEARNER, a framework
based on machine learning that automatically identifies TEs in a
given genome and assigns a classification to them. We present an
implementation of our framework towards LTR retrotransposons, a
particular type of TEs characterized by having long terminal repeats
(LTRs) at their boundaries. We evaluate the predictive performance
of our framework on the well-annotated genomes of Drosophila
melanogaster and Arabidopsis thaliana and we compare our results for
three LTR retrotransposon superfamilies with the results of three
widely used methods for TE identification or classification:
REPEATMASKER, CENSOR and LTRDIGEST. In contrast to these methods,
TE-LEARNER is the first to incorporate machine learning techniques,
outperforming these methods in terms of predictive performance,
while able to learn models and make predictions
efficiently. Moreover, we show that our method was able to identify
TEs that none of the above method could find, and we investigated
TE-LEARNER's predictions which did not correspond to an official
annotation. It turns out that many of these predictions are in fact
strongly homologous to a known TE (<ref xlink:href="#magnet-2018-bid30" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid42" level="1">
      <bodyTitle>A Distributed Frank-Wolfe Framework
for Learning Low-Rank Matrices with the Trace Norm</bodyTitle>
      <p>We consider the problem of learning a high-dimensional but low-rank
matrix from a large-scale dataset distributed over several machines,
where low-rankness is enforced by a convex trace norm constraint. We
propose DFW-Trace, a distributed Frank-Wolfe algorithm which
leverages the low-rank structure of its updates to achieve
efficiency in time, memory and communication usage. The step at the
heart of DFW-Trace is solved approximately using a distributed
version of the power method. We provide a theoretical analysis of
the convergence of DFW-Trace, showing that we can ensure sublinear
convergence in expectation to an optimal solution with few power
iterations per epoch. We implement DFW-Trace in the Apache Spark
distributed programming framework and validate the usefulness of our
approach on synthetic and real data, including the ImageNet dataset
with high-dimensional features extracted from a deep neural
network (<ref xlink:href="#magnet-2018-bid31" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid43" level="1">
      <bodyTitle>Personalized and Private Peer-to-Peer Machine Learning</bodyTitle>
      <p>The rise of connected personal devices together with privacy concerns call for machine learning algorithms capable of leveraging the data of a large number of agents to learn personalized models under strong privacy requirements. In this paper, we introduce an efficient algorithm to address the above problem in a fully decentralized (peer-to-peer) and asynchronous fashion, with provable convergence rate. We show how to make the algorithm differentially private to protect against the disclosure of information about the personal datasets, and formally analyze the trade-off between utility and privacy. Our experiments show that our approach dramatically outperforms previous work in the non-private case, and that under privacy constraints, we can significantly improve over models learned in isolation (<ref xlink:href="#magnet-2018-bid32" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid44" level="1">
      <bodyTitle>Hiding in the Crowd: A Massively Distributed Algorithm for Private Averaging with Malicious Adversaries</bodyTitle>
      <p>The amount of personal data collected in our everyday interactions with connected devices offers great opportunities for innovative services fueled by machine learning, as well as raises serious concerns for the privacy of individuals. In this paper, we propose a massively distributed protocol for a large set of users to privately compute averages over their joint data, which can then be used to learn predictive models. Our protocol can find a solution of arbitrary accuracy, does not rely on a third party and preserves the privacy of users throughout the execution in both the honest-but-curious and malicious adversary models. Specifically, we prove that the information observed by the adversary (the set of malicious users) does not significantly reduce the uncertainty in its prediction of private values compared to its prior belief. The level of privacy protection depends on a quantity related to the Laplacian matrix of the network graph and generally improves with the size of the graph. Furthermore, we design a verification procedure which offers protection against malicious users joining the service with the goal of manipulating the outcome of the algorithm (<ref xlink:href="#magnet-2018-bid33" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid45" level="1">
      <bodyTitle>A Probabilistic Model for Joint Learning of Word Embeddings from Texts and Images</bodyTitle>
      <p>Several recent studies have shown the benefits of combining language and perception to infer word embeddings. These multimodal approaches either simply combine pre-trained textual and visual representations (e.g. features extracted from convolutional neural networks), or use the latter to bias the learning of textual word embeddings. In this work, we propose a novel probabilistic model to formalize how linguistic and perceptual inputs can work in concert to explain the observed word-context pairs in a text corpus. Our approach learns textual and visual representations jointly: latent visual factors couple together a skip-gram model for co-occurrence in linguistic data and a generative latent variable model for visual data. Extensive experimental studies validate the proposed model. Concretely, on the tasks of assessing pairwise word similarity and image/caption retrieval, our approach attains equally competitive or stronger results when compared to other state-of-the-art multimodal models (<ref xlink:href="#magnet-2018-bid34" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid46" level="1">
      <bodyTitle>A Framework for Understanding the Role of Morphology in Universal Dependency Parsing</bodyTitle>
      <p>We present a simple framework for characterizing morphological complexity and how it encodes syntactic information. In particular, we propose a new measure of morpho-syntactic complexity in terms of governor-dependent preferential attachment that explains parsing performance. Through experiments on dependency parsing with data from Universal Dependencies (UD), we show that representations derived from morphological attributes deliver important parsing performance improvements over standard word form embeddings when trained on the same datasets. We also show that the new morpho-syntactic complexity measure is predictive of the gains provided by using morphological attributes over plain forms on parsing scores, making it a tool to distinguish languages using morphology as a syntactic marker from others (<ref xlink:href="#magnet-2018-bid35" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid47" level="1">
      <bodyTitle>Online Reciprocal Recommendation with Theoretical Performance Guarantees</bodyTitle>
      <p>A reciprocal recommendation problem is one where the goal of learning is not just to predict a user's preference towards a passive item (e.g., a book), but to recommend the targeted user on one side another user from the other side such that a mutual interest between the two exists. The problem thus is sharply different from the more traditional items-to-users recommendation, since a good match requires meeting the preferences at both sides. We initiate a rigorous theoretical investigation of the reciprocal recommendation task in a specific framework of sequential learning. We point out general limitations, formulate reasonable assumptions enabling effective learning and, under these assumptions, we design and analyze a computationally efficient algorithm that uncovers mutual likes at a pace comparable to that achieved by a clairvoyant algorithm knowing all user preferences in advance. Finally, we validate our algorithm against synthetic and real-world datasets, showing improved empirical performance over simple baselines (<ref xlink:href="#magnet-2018-bid36" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid48" level="1">
      <bodyTitle>On Similarity Prediction and Pairwise Clustering</bodyTitle>
      <p>We consider the problem of clustering a finite set of items from pairwise similarity information. Unlike what is done in the literature on this subject, we do so in a passive learning setting, and with no specific constraints on the cluster shapes other than their size. We investigate the problem in different settings: i. an online setting, where we provide a tight characterization of the prediction complexity in the mistake bound model, and ii. a standard stochastic batch setting, where we give tight upper and lower bounds on the achievable generalization error. Prediction performance is measured both in terms of the ability to recover the similarity function encoding the hidden clustering and in terms of how well we classify each item within the set. The proposed algorithms are time efficient (<ref xlink:href="#magnet-2018-bid37" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid49" level="1">
      <bodyTitle>A Probabilistic Theory of Supervised Similarity Learning for Pointwise ROC Curve Optimization</bodyTitle>
      <p>The performance of many machine learning techniques depends on the choice of an appropriate similarity or distance measure on the input space. Similarity learning (or metric learning) aims at building such a measure from training data so that observations with the same (resp. different) label are as close (resp. far) as possible. In this paper, similarity learning is investigated from the perspective of pairwise bipartite ranking, where the goal is to rank the elements of a database by decreasing order of the probability that they share the same label with some query data point, based on the similarity scores. A natural performance criterion in this setting is pointwise ROC optimization: maximize the true positive rate under a fixed false positive rate. We study this novel perspective on similarity learning through a rigorous probabilistic framework. The empirical version of the problem gives rise to a constrained optimization formulation involving U-statistics, for which we derive universal learning rates as well as faster rates under a noise assumption on the data distribution. We also address the large-scale setting by analyzing the effect of sampling-based approximations. Our theoretical results are supported by illustrative numerical experiments (<ref xlink:href="#magnet-2018-bid38" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid50" level="1">
      <bodyTitle>Escaping the Curse of Dimensionality in Similarity Learning: Efficient Frank-Wolfe Algorithm and Generalization Bounds</bodyTitle>
      <p>Similarity and metric learning provides a principled approach to construct a task-specific similarity from weakly supervised data. However, these methods are subject to the curse of dimensionality: as the number of features grows large, poor generalization is to be expected and training becomes intractable due to high computational and memory costs. In this paper, we propose a similarity learning method that can efficiently deal with high-dimensional sparse data. This is achieved through a parameterization of similarity functions by convex combinations of sparse rank-one matrices, together with the use of a greedy approximate Frank-Wolfe algorithm which provides an efficient way to control the number of active features. We show that the convergence rate of the algorithm, as well as its time and memory complexity, are independent of the data dimension. We further provide a theoretical justification of our modeling choices through an analysis of the generalization error, which depends logarithmically on the sparsity of the solution rather than on the number of features. Our experiments on datasets with up to one million features demonstrate the ability of our approach to generalize well despite the high dimensionality as well as its superiority compared to several competing methods (<ref xlink:href="#magnet-2018-bid39" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
    <subsection id="uid51" level="1">
      <bodyTitle>Nonstochastic Bandits with Composite Anonymous Feedback</bodyTitle>
      <p>We investigate a nonstochastic bandit setting in which the loss of an action is not immediately charged to the player, but rather spread over at most d consecutive steps in an adversarial way. This implies that the instantaneous loss observed by the player at the end of each round is a sum of as many as d loss components of previously played actions. Hence, unlike the standard bandit setting with delayed feedback, here the player cannot observe the individual delayed losses, but only their sum. Our main contribution is a general reduction transforming a standard bandit algorithm into one that can operate in this harder setting. We also show how the regret of the transformed algorithm can be bounded in terms of the regret of the original algorithm. Our reduction cannot be improved in general: we prove a lower bound on the regret of any bandit algorithm in this setting that matches (up to log factors) the upper bound obtained via our reduction. Finally, we show how our reduction can be extended to more complex bandit settings, such as combinatorial linear bandits and online bandit convex optimization (<ref xlink:href="#magnet-2018-bid40" location="biblio" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest"/>).
</p>
    </subsection>
  </resultats>
  <contrats id="uid52">
    <bodyTitle>Bilateral Contracts and Grants with Industry</bodyTitle>
    <subsection id="uid53" level="1">
      <bodyTitle>Coreference resolution</bodyTitle>
      <p>Along a collaboration with Orange, we developed a Natural Language Processing
library for co-reference resolution. The library is based on a
previous work (CorTeX) and was extended in several ways. It
handles the French language, it includes new features based on
vectorial representations of words (word embeddings) and it is
more scalable. <span class="smallcap" align="left">Pascal Denis</span> is the local PI at Inria of this project.
</p>
    </subsection>
    <subsection id="uid54" level="1">
      <bodyTitle>Privacy preserving data mining for Mobility Data</bodyTitle>
      <p><span class="smallcap" align="left">Jan Ramon</span> is the local PI at Inria for the ADEME-MUST project
(Méthodologie d'exploitation des données d'usage des véhicules
et d'identification de nouveaux services pour les usagers et les
territoires). We study machine learning and data mining methods for
knowledge discovery from mobility data, which are time-stamped
signals collected from cars, for example, GPS locations,
accelerations and fuel consumption. We aim to discover knowledge
that helps us to address important questions in the transportation
system such as road safety, traffic congestion, parking,
ride-sharing, pollution and energy consumption. As the mobility data
contains a lot of personal information, for instance, driving styles
and locations of the users, we hence also study methods that allow
the users to keep their personal data and only exchange part of them
to collaboratively derive the knowledge.</p>
      <p>The project has four partners, including, Xee company, CEREMA,
i-Trans and Inria. The Xee company is responsible for recruiting
drivers and collecting the data. CEREMA and i-Trans function as
domain experts who help us to form the questions and verify the
analytical results. <span class="smallcap" align="left">Magnet</span> is responsible for developing and
applying data mining methods for analyzing the data. The developed
methods and the discovered knowledge from the project will be
transferred to Metropole Lille and ADEME.
</p>
    </subsection>
    <subsection id="uid55" level="1">
      <bodyTitle>Predictive justice</bodyTitle>
      <p>Claim assistance is a French company that develops assistance for
conflict resolution. The main service is
RefundMyTicket <footnote id="uid56" id-text="1"><ref xlink:href="https://www.refundmyticket.net" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>www.<allowbreak/>refundmyticket.<allowbreak/>net</ref></footnote>. In
the general project of partial automation of analysis of complains,
we have provided consulting and supervision. The general approach
was to be able to analyze, parse and reason on legal texts. We have
developed strategies based on natural language processing in the
specific domain of legal texts. Techniques include learning
representation and structured prediction among others.
</p>
    </subsection>
  </contrats>
  <partenariat id="uid57">
    <bodyTitle>Partnerships and Cooperations</bodyTitle>
    <subsection id="uid58" level="1">
      <bodyTitle>Regional Initiatives</bodyTitle>
      <p>We conducted research in collaboration with J. Senechal from the
department of law in Lille University. We are interested in studying
the impact of technological choices regarding computation models in
the perspective of the GDPR.</p>
      <p>We strengthened our partnership with the linguistic laboratory STL
in Lille university. We have welcomed Bert Cappelle for a stay
(delegation) in the group. The topic of this collaboration was to
study modal verbs and the translation of the notion of
compositionality when applied to vectorial representation of words.</p>
      <p>We initiated a collaboration with cognitive
scientists (Angèle Brunellière and Jérémie Jozefowiez) from the
psychology department, which resulted in a submission to a
multidisciplinary Huma-Num project, to be funded by the Réseau
National des Maisons des Sciences de l'Homme (RNMSH).</p>
      <p>We started working with Christopher Fletcher (CNRS) from the History
department.</p>
      <p>These collaborations heavily rely on our work on distributional
semantics and word embeddings to provide new insights into these
different fields, hence also on the Mangoes toolkit developed in the
team.</p>
      <p>We participate to the <i>Data Advanced data science and
technologies</i> project (CPER Data). This project is organized
following three axes: internet of things, data science, high
performance computing. <span class="smallcap" align="left">Magnet</span> is involved in the data science axis
to develop machine learning algorithms for big data, structured data
and heterogeneous data. The project MyLocalInfo is an open API for
privacy-friendly collaborative computing in the internet of things.
</p>
    </subsection>
    <subsection id="uid59" level="1">
      <bodyTitle>National Initiatives</bodyTitle>
      <subsection id="uid60" level="2">
        <bodyTitle>ANR Pamela (2016-2020)</bodyTitle>
        <p><b>Participants</b>: <span class="smallcap" align="left">Marc Tommasi</span> [correspondent], <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Rémi Gilleron</span>, <span class="smallcap" align="left">Jan Ramon</span>, <span class="smallcap" align="left">Mahsa Asadi</span></p>
        <p>The Pamela project aims at developing machine learning theories and algorithms in order to learn local and personalized models from data distributed over networked infrastructures. Our project seeks to provide first answers to modern information systems built by interconnecting many personal devices holding private user data in the search of personalized suggestions and recommendations. More precisely, we will focus on learning in a collaborative way with the help of neighbors in a network. We aim to lay the first blocks of a scientific foundation for these new types of systems, in effect moving from graphs of data to graphs of data and learned models. We argue that this shift is necessary in order to address the new constraints arising from the decentralization of information that is inherent to the emergence of big data. We will in particular focus on the question of learning under communication and privacy constraints. A significant asset of the project is the quality of its industrial partners, Snips and Mediego, who bring in their expertise in privacy protection and distributed computing as well as use cases and datasets. They will contribute to translate this fundamental research effort into concrete outcomes by developing personalized and privacy-aware assistants able to provide contextualized recommendations on small devices and smartphones.
<ref xlink:href="https://project.inria.fr/pamela/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>project.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>pamela/</ref>.</p>
      </subsection>
      <subsection id="uid61" level="2">
        <bodyTitle>ANR JCJC GRASP (2016-2020)</bodyTitle>
        <p><b>Participants</b>: <span class="smallcap" align="left">Pascal Denis</span> [correspondent], <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Rémi Gilleron</span>, <span class="smallcap" align="left">Mikaela Keller</span>, <span class="smallcap" align="left">Marc Tommasi</span></p>
        <p>The GRASP project aims at designing new graph-based Machine Learning algorithms that are better tailored to Natural Language Processing structured output problems. Focusing on semi-supervised learning scenarios, we will extend current graph-based learning approaches along two main directions: (i) the use of structured outputs during inference, and (ii) a graph construction mechanism that is more dependent on the task objective and more closely related to label inference. Combined, these two research strands will provide an important step towards delivering more adaptive (to new domains and languages), more accurate, and ultimately more useful language technologies. We will target semantic and pragmatic tasks such as coreference resolution, temporal chronology prediction, and discourse parsing for which proper Machine Learning solutions are still lacking.
<ref xlink:href="https://project.inria.fr/grasp/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>project.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>grasp/</ref>.</p>
      </subsection>
      <subsection id="uid62" level="2">
        <bodyTitle>ANR DEEP-Privacy (2019-2023)</bodyTitle>
        <p><b>Participants</b>: <span class="smallcap" align="left">Marc Tommasi</span> [correspondent], <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Pascal Denis</span>, <span class="smallcap" align="left">Jan Ramon</span>, <span class="smallcap" align="left">Brij Srivastava</span></p>
        <p>DEEP-PRIVACY proposes a new paradigm based on a distributed, personalized, and privacy-preserving approach for speech processing, with a focus on machine learning algorithms for speech recognition. To this end, we propose to rely on a hybrid approach: the device of each user does not share its raw speech data and runs some private computations locally, while some cross-user computations are done by communicating through a server (or a peer-to-peer network). To satisfy privacy requirements at the acoustic level, the information communicated to the server should not expose sensitive speaker information.</p>
      </subsection>
      <subsection id="uid63" level="2">
        <bodyTitle>ANR-NFS REM (2016-2020)</bodyTitle>
        <p><b>Participants</b>: <span class="smallcap" align="left">Pascal Denis</span> [correspondent], <span class="smallcap" align="left">Bo Li</span></p>
        <p>With colleagues from the linguistics departments at Lille 3 and Neuchâtel (Switzerland), <span class="smallcap" align="left">Pascal Denis</span> is a member of another ANR project (REM), funded through the bilateral ANR-NFS Scheme. This project, co-headed by I. Depreatere (Lille 3) and M. Hilpert (Neufchâtel), proposes to reconsider the analysis of English modal constructions from a multidisciplinary perspective, combining insights from theoretical, psycho-linguistic, and computational approaches.</p>
      </subsection>
      <subsection id="uid64" level="2">
        <bodyTitle>EFL (2010-2020)</bodyTitle>
        <p><span class="smallcap" align="left">Pascal Denis</span> is an associate member of the Laboratoire d'Excellence <i>Empirical Foundations of Linguistics</i> (EFL), <ref xlink:href="http://www.labex-efl.org/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>www.<allowbreak/>labex-efl.<allowbreak/>org/</ref>.</p>
      </subsection>
    </subsection>
    <subsection id="uid65" level="1">
      <bodyTitle>European Initiatives</bodyTitle>
      <subsection id="uid66" level="2">
        <bodyTitle>FP7 &amp; H2020 Projects</bodyTitle>
        <sanspuceslist>
          <li id="uid67">
            <p noindent="true">Program: H2020 ICT-29-2018 (RIA)</p>
          </li>
          <li id="uid68">
            <p noindent="true">Project acronym: COMPRISE</p>
          </li>
          <li id="uid69">
            <p noindent="true">Project title: Cost-effective, Multilingual, Privacy-driven voice-enabled Services</p>
          </li>
          <li id="uid70">
            <p noindent="true">Duration: Dec 2018- Nov 2021</p>
          </li>
          <li id="uid71">
            <p noindent="true">Coordinator: Emmanuel Vincent</p>
          </li>
          <li id="uid72">
            <p noindent="true">Other partners: Inria Multispeech, Ascora GmbH, Netfective Technology SA, Rooter Analysis SL, Tilde SIA, University of Saarland</p>
          </li>
          <li id="uid73">
            <p noindent="true">Participants: <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Marc Tommasi</span>, <span class="smallcap" align="left">Brij Srivastava</span></p>
          </li>
          <li id="uid74">
            <p noindent="true">Abstract: COMPRISE will define a fully private-by-design methodology and tools that will reduce the cost and increase the inclusiveness of voice interaction technologies.</p>
          </li>
        </sanspuceslist>
      </subsection>
      <subsection id="uid75" level="2">
        <bodyTitle>Collaborations in European Programs, Except FP7 &amp; H2020</bodyTitle>
        <subsection id="uid76" level="3">
          <bodyTitle>TextLink (2014-2018)</bodyTitle>
          <sanspuceslist>
            <li id="uid77">
              <p noindent="true">Program: COST Action</p>
            </li>
            <li id="uid78">
              <p noindent="true">Project acronym: TextLink</p>
            </li>
            <li id="uid79">
              <p noindent="true">Project title: Structuring Discourse in Multilingual Europe</p>
            </li>
            <li id="uid80">
              <p noindent="true">Duration: Apr. 2014 - Apr. 2018</p>
            </li>
            <li id="uid81">
              <p noindent="true">Coordinator: Prof. Liesbeth Degand, Université Catholique de Louvain, Belgium. <span class="smallcap" align="left">Pascal Denis</span> is member of the Tools group.</p>
            </li>
            <li id="uid82">
              <p noindent="true">Other partners: 26 EU countries and 3 international partner countries (Argentina, Brazil, Canada)</p>
            </li>
            <li id="uid83">
              <p noindent="true">The Action will facilitate European multilingualism by (1) identifying and creating a portal into such resources within Europe - including annotation tools, search tools, and discourse-annotated corpora; (2) delineating the dimensions and properties of discourse annotation across corpora; (3) organizing these properties into a sharable taxonomy; (4) encouraging the use of this taxonomy in subsequent discourse annotation and in cross-lingual search and studies of devices that relate and structure discourse; and (5) promoting use of the portal, its resources and sharable taxonomy. TextLink will enhance the experience and performance of human translators, lexicographers, language technology and language learners alike.</p>
            </li>
          </sanspuceslist>
        </subsection>
      </subsection>
    </subsection>
    <subsection id="uid84" level="1">
      <bodyTitle>International Initiatives</bodyTitle>
      <subsection id="uid85" level="2">
        <bodyTitle>Inria International Labs</bodyTitle>
        <p>
          <b>Inria@SiliconValley</b>
        </p>
        <p noindent="true">Associate Team involved in the International Lab:</p>
        <subsection id="uid86" level="3">
          <bodyTitle>
            <ref xlink:href="https://team.inria.fr/lego/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">LEGO </ref>
          </bodyTitle>
          <sanspuceslist>
            <li id="uid87">
              <p noindent="true">Title: LEarning GOod representations for natural language processing</p>
            </li>
            <li id="uid88">
              <p noindent="true">International Partner (Institution - Laboratory - Researcher):</p>
              <sanspuceslist>
                <li id="uid89">
                  <p noindent="true">USC (United States), Prof. Fei Sha.</p>
                </li>
              </sanspuceslist>
            </li>
            <li id="uid90">
              <p noindent="true">Start year: 2016</p>
            </li>
            <li id="uid91">
              <p noindent="true">See also: <ref xlink:href="https://team.inria.fr/lego/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>team.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>lego/</ref></p>
            </li>
            <li id="uid92">
              <p noindent="true">LEGO lies in the intersection of Machine Learning and Natural Language Processing (NLP). Its goal is to address the following challenges: what are the right representations for structured data and how to learn them automatically, and how to apply such representations to complex and structured prediction tasks in NLP? In recent years, continuous vectorial embeddings learned from massive unannotated corpora have been increasingly popular, but they remain far too limited to capture the complexity of text data as they are task-agnostic and fall short of modeling complex structures in languages. LEGO strongly relies on the complementary expertise of the two partners in areas such as representation/similarity learning, structured prediction, graph-based learning, and statistical NLP to offer a novel alternative to existing techniques. Specifically, we will investigate the following three research directions: (a) optimize the embeddings based on annotations so as to minimize structured prediction errors, (b) generate embeddings from rich language contexts represented as graphs, and (c) automatically adapt the context graph to the task/dataset of interest by learning a similarity between nodes to appropriately weigh the edges of the graph. By exploring these complementary research strands, we intend to push the state-of-the-art in several core NLP problems, such as dependency parsing, coreference resolution and discourse parsing.</p>
            </li>
          </sanspuceslist>
        </subsection>
      </subsection>
      <subsection id="uid93" level="2">
        <bodyTitle>Inria Associate Teams Not Involved in an Inria International Labs</bodyTitle>
        <sanspuceslist>
          <li id="uid94">
            <p noindent="true">North-European Associate Team PAD-ML: Privacy-Aware Distributed Machine Learning.</p>
          </li>
          <li id="uid95">
            <p noindent="true">International Partner: the PPDA team at the Alan Turing Institute.</p>
          </li>
          <li id="uid96">
            <p noindent="true">Start year: 2018</p>
          </li>
          <li id="uid97">
            <p noindent="true">In the context of increasing legislation on data protection (e.g., the recent GDPR), an important challenge is to develop privacy-preserving algorithms to learn from datasets distributed across multiple data owners who do not want to share their data. The goal of this joint team is to devise novel privacy-preserving, distributed machine learning algorithms and to assess their performance and guarantees in both theoretical and practical terms.</p>
          </li>
        </sanspuceslist>
      </subsection>
    </subsection>
    <subsection id="uid98" level="1">
      <bodyTitle>International Research Visitors</bodyTitle>
      <subsection id="uid99" level="2">
        <bodyTitle>Visits of International Scientists</bodyTitle>
        <simplelist>
          <li id="uid100">
            <p noindent="true">Tejas Kulkarni (University of Warwick) visited the team from May to August 2018 to work with <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Marc Tommasi</span> and <span class="smallcap" align="left">Jan Ramon</span> on privacy-preserving computation of <formula type="inline"><math xmlns="http://www.w3.org/1998/Math/MathML" overflow="scroll"><mi>U</mi></math></formula>-statistics.</p>
          </li>
          <li id="uid101">
            <p noindent="true">Larisa Soldatova (Brunel University) visited the team in June 2018 to work with <span class="smallcap" align="left">Jan Ramon</span> on probabilistic reasoning for biomedical applications.</p>
          </li>
          <li id="uid102">
            <p noindent="true">Raouf Kerkouche (Inria Privatics) visited the team for 2 weeks in July 2018 to work with <span class="smallcap" align="left">Aurélien Bellet</span> and <span class="smallcap" align="left">Marc Tommasi</span> on federated and decentralized learning from medical data.</p>
          </li>
          <li id="uid103">
            <p noindent="true">Guillaume Rabusseau (Université de Montréal) visited the team for 1 week in July 2018 to work with <span class="smallcap" align="left">Aurélien Bellet</span> and <span class="smallcap" align="left">Marc Tommasi</span> on multi-task distributed spectral learning.</p>
          </li>
          <li id="uid104">
            <p noindent="true">Daphner Ezer, Adrià Gascón, Matt Kusner, Brooks Paige (all from Alan Turing Institute) and Hamed Haddadi (Imperial College London) visited the team for 2 days in October 2018 for the kick-off of the PAD-ML associate team.</p>
          </li>
        </simplelist>
        <p>Several international researchers have also been invited to give a talk at the MAGNET seminar:</p>
        <simplelist>
          <li id="uid105">
            <p noindent="true">D. Hovy (Bocconi Univ.): Retrofit Everything: Injecting External Knowledge into Neural Networks to Gain Insights from Big Data.</p>
          </li>
          <li id="uid106">
            <p noindent="true">A. Trask (OpenMined): OpenMined - Building Tools for Safe AI.</p>
          </li>
          <li id="uid107">
            <p noindent="true">C. Biemann (Univ. Hamburg): Adaptive Interpretable Language Technology.</p>
          </li>
          <li id="uid108">
            <p noindent="true">W. Daelemans (Univ. Antwerp): Profiling authors from social media texts.</p>
          </li>
        </simplelist>
        <subsection id="uid109" level="3">
          <bodyTitle>Internships</bodyTitle>
          <simplelist>
            <li id="uid110">
              <p noindent="true">Igor Axinti explored several ways to compare word embeddings and
studied the minimal corpus size for the comparison to be
meaningful. He applied some of his findings to comparing two corpus in
middle french from the 15th century, one originating from London and
the other from Flanders. He produced a querying interface to allow
Christopher Fletcher (IRHiS), who provided the data, explore and
compare the embeddings spaces.</p>
            </li>
            <li id="uid111">
              <p noindent="true">Nicolas Crosetti (joint internship with Joachim Niehren and
Florent Cappelli, Links) worked on dependency-weighted aggregation,
i.e., aggregation where the elements to aggregate are weighted
according to the extent where they correspond to independent
observations.</p>
            </li>
            <li id="uid112">
              <p noindent="true">Arthur d'Azemar worked on decentralized recommender systems in
collaboration with the WIDE team in Inria Rennes (François
Taïani). Arthur has applied metric learning techniques in order to
learn a K-nn graph for personalized and adaptive user-based
recommendations.</p>
            </li>
            <li id="uid113">
              <p noindent="true">Antoine Capriski worked on the analysis of word semantic change
in political texts in collaboration with Caroline Le Pennec (UC
Berkeley). He used the techniques of word embeddings to analyze of
corpus of political manifestos from the French general elections for
the period 1958-1993.</p>
            </li>
            <li id="uid114">
              <p noindent="true">Most of the works on machine learning and privacy make the
assumption that learners are honest but curious. Alexandre Huat
worked on making protocols for private machine learning more robust
again malicious attacks.</p>
            </li>
          </simplelist>
        </subsection>
      </subsection>
      <subsection id="uid115" level="2">
        <bodyTitle>Visits to International Teams</bodyTitle>
        <subsection id="uid116" level="3">
          <bodyTitle>Research Stays Abroad</bodyTitle>
          <simplelist>
            <li id="uid117">
              <p noindent="true"><span class="smallcap" align="left">Fabio Vitale</span> is on leave at Department of Computer Science of Sapienza University (Rome, Italy) in the Algorithms Randomization Computation group with Prof. Alessandro Panconesi and Prof. Flavio Chierichetti. His current work on machine learning in graphs follows three directions:</p>
              <simplelist>
                <li id="uid118">
                  <p noindent="true">designing new online reciprocal recommenders analyzing their performance both in theory and in practice,</p>
                </li>
                <li id="uid119">
                  <p noindent="true">clustering a finite set of items from pairwise similarity information in different learning settings,</p>
                </li>
                <li id="uid120">
                  <p noindent="true">introducing a new online learning framework encompassing several problems where the environment changes over time, and an efficient and very scalable unifying approach to solve the related general learning problem.</p>
                </li>
              </simplelist>
              <p>Current (and unfinished) ongoing research also includes the following topics: low-stretch spanning trees, active learning in correlation clustering problems, hierarchical clustering.</p>
            </li>
            <li id="uid121">
              <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> visited the Alan Turing Institute (London) and Amazon Research Cambridge for 1 week in February 2018. He worked with Adrià Gascón and Borja Balle on privacy-preserving machine learning.</p>
            </li>
          </simplelist>
        </subsection>
      </subsection>
    </subsection>
  </partenariat>
  <diffusion id="uid122">
    <bodyTitle>Dissemination</bodyTitle>
    <subsection id="uid123" level="1">
      <bodyTitle>Promoting Scientific Activities</bodyTitle>
      <subsection id="uid124" level="2">
        <bodyTitle>Scientific Events Organisation</bodyTitle>
        <subsection id="uid125" level="3">
          <bodyTitle>Member of the Organizing Committees</bodyTitle>
          <simplelist>
            <li id="uid126">
              <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> was a member of the organization committee of the PPML workshop at NeurIPS'18. <footnote id="uid127" id-text="2"><ref xlink:href="https://neurips.cc/Conferences/2018/Schedule?showEvent=10934" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>neurips.<allowbreak/>cc/<allowbreak/>Conferences/<allowbreak/>2018/<allowbreak/>Schedule?showEvent=10934</ref></footnote> The workshop was on Privacy Preserving Machine Learning and had among its invited speakers Shafi Goldwasser (Gödel and Turing Prize), Adam Smith (Gödel Prize).</p>
            </li>
            <li id="uid128">
              <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> co-organized the kick-off workshop of the associated team PAD-ML with the Alan Turing Institute. <footnote id="uid129" id-text="3"><ref xlink:href="https://team.inria.fr/magnet/workshop-on-privacy-aware-distributed-machine-learning/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>team.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>magnet/<allowbreak/>workshop-on-privacy-aware-distributed-machine-learning/</ref></footnote> The workshop was held at Inria Lille and featured speakers from <span class="smallcap" align="left">Magnet</span>  and the Alan Turing Institute.</p>
            </li>
          </simplelist>
        </subsection>
      </subsection>
      <subsection id="uid130" level="2">
        <bodyTitle>Scientific Events Selection</bodyTitle>
        <subsection id="uid131" level="3">
          <bodyTitle>Member of the Conference Program Committees</bodyTitle>
          <simplelist>
            <li id="uid132">
              <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> served as PC member for AISTATS'19, ICML'18, NIPS'18,
IJCAI'18 Sister Conference, PiMLAI workshop at ICML'18, and
CAP'18.</p>
            </li>
            <li id="uid133">
              <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> served as PC member for ACL'18, CONLL'18, EMNLP'18,
NAACL'18, NIPS'18, IJCAI-ECAI'18 (Senior PC), CRAC Workshop at NAACL'18.</p>
            </li>
            <li id="uid134">
              <p noindent="true"><span class="smallcap" align="left">Marc Tommasi</span> served as PC member for AAAI'18, ICML'18, CAP'18, IJCAI'18 (Senior PC chair), AISTATS'18, NIPS'18.</p>
            </li>
            <li id="uid135">
              <p noindent="true"><span class="smallcap" align="left">Jan Ramon</span> served as PC member for AAAI'19, AISTATS'19, IEEE-BigData'18, CIKM'18, DS'18, ECML/PKDD'18, EKAW'18, IEEE-ICDM'18, ICML'18, ILP'18, LOD'18, MLG'18, NIPS'18, SDM'18, TDLSG'18.</p>
            </li>
            <li id="uid136">
              <p noindent="true"><span class="smallcap" align="left">Mikaela Keller</span> served as PC member for ICML'18, CAP'18.</p>
            </li>
            <li id="uid137">
              <p noindent="true"><span class="smallcap" align="left">Rémi Gilleron</span> served as PC member for NIPS'18, CAP'18, AISTATS'19 and ICLR'19.</p>
            </li>
          </simplelist>
        </subsection>
      </subsection>
      <subsection id="uid138" level="2">
        <bodyTitle>Journal</bodyTitle>
        <subsection id="uid139" level="3">
          <bodyTitle>Reviewer - Reviewing Activities</bodyTitle>
          <simplelist>
            <li id="uid140">
              <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> was reviewer for Machine Learning Journal and IEEE/ACM
Transactions on Networking.</p>
            </li>
            <li id="uid141">
              <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> was reviewer for Computational Linguistics, IJCAI-ECAI
Surveys, and Language Resources and Evaluation.</p>
            </li>
            <li id="uid142">
              <p noindent="true"><span class="smallcap" align="left">Jan Ramon</span> was member of the editorial boards of Machine Learning Journal (MLJ) and Data Mining and Knowledge Discovery (DMKD). <span class="smallcap" align="left">Jan Ramon</span> was reviewer for among others JMLR, TPAMI, JIIS.</p>
            </li>
          </simplelist>
        </subsection>
      </subsection>
      <subsection id="uid143" level="2">
        <bodyTitle>Invited Talks</bodyTitle>
        <simplelist>
          <li id="uid144">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> gave invited talks at the EPFL-Inria 2018 workshop <footnote id="uid145" id-text="4"><ref xlink:href="https://project.inria.fr/epfl-Inria/workshops/workshop-2018/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>project.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>epfl-Inria/<allowbreak/>workshops/<allowbreak/>workshop-2018/</ref></footnote> and the Journées de Statistique 2018 (session SSFAM). <footnote id="uid146" id-text="5"><ref xlink:href="http://jds2018.sfds.asso.fr/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>jds2018.<allowbreak/>sfds.<allowbreak/>asso.<allowbreak/>fr/</ref></footnote></p>
          </li>
          <li id="uid147">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> was invited to talk at the seminars of Inria WIDE,
Télécom ParisTech, Statistics Seminar of Paris 6/7, CMLA (ENS
Paris Saclay) and Naver Labs Europe.</p>
          </li>
          <li id="uid148">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> gave an invited talk at the Séminaire Langage, SCALab,
Université de Lille, 26/01/18.</p>
          </li>
        </simplelist>
      </subsection>
      <subsection id="uid149" level="2">
        <bodyTitle>Scientific Expertise</bodyTitle>
        <simplelist>
          <li id="uid150">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> was a member of the jury for the Gilles-Kahn PhD award of the French Society of Computer
Science (SIF), sponsored by the French Academy of Sciences. <footnote id="uid151" id-text="6"><ref xlink:href="https://www.societe-informatique-de-france.fr/recherche/prix-de-these-gilles-kahn/" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>www.<allowbreak/>societe-informatique-de-france.<allowbreak/>fr/<allowbreak/>recherche/<allowbreak/>prix-de-these-gilles-kahn/</ref></footnote></p>
          </li>
          <li id="uid152">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> acted as external reviewer for the French National Research Agency (ANR), track “Projets de
Recherche Collaborative – International”.</p>
          </li>
          <li id="uid153">
            <p noindent="true"><span class="smallcap" align="left">Jan Ramon</span> was an external reviewer for the Swiss National Science Foundation (SNF).</p>
          </li>
          <li id="uid154">
            <p noindent="true"><span class="smallcap" align="left">Jan Ramon</span> was an external reviewer for the Vienna Science and Technology Fund (WWTF).</p>
          </li>
          <li id="uid155">
            <p noindent="true"><span class="smallcap" align="left">Jan Ramon</span> acted as an expert for the H2020 CoE and IMI programs.</p>
          </li>
        </simplelist>
      </subsection>
      <subsection id="uid156" level="2">
        <bodyTitle>Research Administration</bodyTitle>
        <simplelist>
          <li id="uid157">
            <p noindent="true"><span class="smallcap" align="left">Mikaela Keller</span> is member of the Conseil du laboratoire CRIStAL.</p>
          </li>
          <li id="uid158">
            <p noindent="true"><span class="smallcap" align="left">Fabien Torre</span> is member of the bureau du Conseil National des Universités (section 27).</p>
          </li>
          <li id="uid159">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> served as a member of the CNRS Pre-GDR NLP Group.</p>
          </li>
          <li id="uid160">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> was elected to Comité National du CNRS, section 34 (Sciences du
Langage).</p>
          </li>
        </simplelist>
      </subsection>
    </subsection>
    <subsection id="uid161" level="1">
      <bodyTitle>Teaching - Supervision - Juries</bodyTitle>
      <subsection id="uid162" level="2">
        <bodyTitle>Teaching</bodyTitle>
        <sanspuceslist>
          <li id="uid163">
            <p noindent="true">Licence SHS: <span class="smallcap" align="left">Joël Legrand</span>, Traitement de textes et tableur, 10h, L1, Université Lille.</p>
          </li>
          <li id="uid164">
            <p noindent="true">Licence SHS: <span class="smallcap" align="left">Marc Tommasi</span>, Langages du Web, 24h, L2, Université Lille.</p>
          </li>
          <li id="uid165">
            <p noindent="true">Licence MIASHS: <span class="smallcap" align="left">Mikaela Keller</span>, Python 1, 40h, L1, Université Lille.</p>
          </li>
          <li id="uid166">
            <p noindent="true">Licence MIASHS: <span class="smallcap" align="left">Marc Tommasi</span>, Codage et représentation de l'information, 48h, L1, Université Lille.</p>
          </li>
          <li id="uid167">
            <p noindent="true">Licence MIASHS: <span class="smallcap" align="left">Mikaela Keller</span>, Codage et représentation de l'information, 42h, L1, Université Lille.</p>
          </li>
          <li id="uid168">
            <p noindent="true">Licence SoQ (SHS): <span class="smallcap" align="left">Mikaela Keller</span>, Algorithmique de graphes, 24h, L3, Université Lille.</p>
          </li>
          <li id="uid169">
            <p noindent="true">Licence <span class="smallcap" align="left">Marc Tommasi</span> C2i 12h, Université Lille.</p>
          </li>
          <li id="uid170">
            <p noindent="true">Licence <span class="smallcap" align="left">Marc Tommasi</span> Humanités numériques - Découvrir et faire découvrir la programmation, 20h, Université Lille/</p>
          </li>
          <li id="uid171">
            <p noindent="true">Master MIASHS: <span class="smallcap" align="left">Mikaela Keller</span>, Algorithmes fondamentaux de la fouille de données, 60h, M1, Université Lille.</p>
          </li>
          <li id="uid172">
            <p noindent="true">Master MIASHS: <span class="smallcap" align="left">Joël Legrand</span>, Apprentissage et émergence de comportements, 30h, M2, Université Lille.</p>
          </li>
          <li id="uid173">
            <p noindent="true">Master Data Analysis &amp; Decision Making: <span class="smallcap" align="left">Aurélien Bellet</span>, Machine Learning, 12h, Ecole Centrale de Lille.</p>
          </li>
          <li id="uid174">
            <p noindent="true">Master / Master Spécialisé Big Data: <span class="smallcap" align="left">Aurélien Bellet</span>, Advanced Machine Learning, 15h, Télécom ParisTech.</p>
          </li>
          <li id="uid175">
            <p noindent="true">Formation continue (Certificat d’Études Spécialisées Data
Scientist): <span class="smallcap" align="left">Aurélien Bellet</span>, Supervised Learning and Support Vector Machines,
17.5h, Télécom ParisTech.</p>
          </li>
          <li id="uid176">
            <p noindent="true">Master Informatique: <span class="smallcap" align="left">Pascal Denis</span>, Fondements de l'Apprentissage
Automatique, 46h, M1, Université de Lille.</p>
          </li>
        </sanspuceslist>
      </subsection>
      <subsection id="uid177" level="2">
        <bodyTitle>Supervision</bodyTitle>
        <sanspuceslist>
          <li id="uid178">
            <p noindent="true">Postdoc: <span class="smallcap" align="left">Melissa Ailem</span>, InriaSiliconValley postdoctoral grant,
supervised by <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Marc Tommasi</span>, <span class="smallcap" align="left">Pascal Denis</span> and <span class="smallcap" align="left">Fei Sha</span> (University of
Southern California).</p>
          </li>
          <li id="uid179">
            <p noindent="true">Postdoc: <span class="smallcap" align="left">Bo Li</span>, supervised by <span class="smallcap" align="left">Pascal Denis</span> on ANR REM, Model Sense Disambiguation, since December 2017.</p>
          </li>
          <li id="uid180">
            <p noindent="true">PhD: <span class="smallcap" align="left">Géraud Le Falher</span>, Characterizing edges in signed and vector-valued graphs. April 16th 2018, <span class="smallcap" align="left">Marc Tommasi</span> and <span class="smallcap" align="left">Fabio Vitale</span> and <span class="smallcap" align="left">Claudio Gentile</span>.</p>
          </li>
          <li id="uid181">
            <p noindent="true">Phd: <span class="smallcap" align="left">Ashraf M. Kibriya</span>, Mining Frequent Patterns in Large Networks, June 2018, <span class="smallcap" align="left">Jan Ramon</span>.</p>
          </li>
          <li id="uid182">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Mathieu Dehouck</span>, Graph-based Learning for Multi-lingual and Multi-domain Dependency Parsing, since Oct 2015, <span class="smallcap" align="left">Pascal Denis</span> and <span class="smallcap" align="left">Marc Tommasi</span>.</p>
          </li>
          <li id="uid183">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Onkar Pandit</span>, Graph-based Semi-supervised Linguistic Structure Prediction, since Dec. 2017, <span class="smallcap" align="left">Pascal Denis</span>, <span class="smallcap" align="left">Marc Tommasi</span> and <span class="smallcap" align="left">Liva Ralaivola</span> (University of Marseille).</p>
          </li>
          <li id="uid184">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Mariana Vargas Vieyra</span>, Adaptive Graph Learning with Applications to Natural Language Processing, since Jan. 2018. <span class="smallcap" align="left">Pascal Denis</span> and <span class="smallcap" align="left">Aurélien Bellet</span> and <span class="smallcap" align="left">Marc Tommasi</span>.</p>
          </li>
          <li id="uid185">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Brij Srivastava</span>, Representation Learning for Privacy-Preserving Speech Recognition, since Oct 2018 <span class="smallcap" align="left">Aurélien Bellet</span> and <span class="smallcap" align="left">Marc Tommasi</span> and <span class="smallcap" align="left">Emmanuel Vincent</span>.</p>
          </li>
          <li id="uid186">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Mahsa Asadi</span>, On Decentralized Machine Learning, since Oct 2018. <span class="smallcap" align="left">Aurélien Bellet</span> and <span class="smallcap" align="left">Marc Tommasi</span>.</p>
          </li>
          <li id="uid187">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Nicolas Crosetti</span>, Privacy Risks of Aggregates in Data Centric-Workflows, since Oct 2018. <span class="smallcap" align="left">Florent Capelli</span> and <span class="smallcap" align="left">Sophie Tison</span> and <span class="smallcap" align="left">Joachim Niehren</span> and <span class="smallcap" align="left">Jan Ramon</span>.</p>
          </li>
          <li id="uid188">
            <p noindent="true">PhD in progress: <span class="smallcap" align="left">Robin Vogel</span>, Learning to rank by similarity and performance optimization in biometric identification, since 2017 (CIFRE thesis with IDEMIA and Télécom ParisTech). <span class="smallcap" align="left">Aurélien Bellet</span>, <span class="smallcap" align="left">Stéphan Clémençon</span> and <span class="smallcap" align="left">Anne Sabourin</span>.</p>
          </li>
        </sanspuceslist>
      </subsection>
      <subsection id="uid189" level="2">
        <bodyTitle>Juries</bodyTitle>
        <simplelist>
          <li id="uid190">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> was member of the PhD jury of Guillaume Papa (Télécom ParisTech), Wenjie Zheng (Sorbonne Université), Michael Blot (Sorbonne Université).</p>
          </li>
          <li id="uid191">
            <p noindent="true"><span class="smallcap" align="left">Marc Tommasi</span> was member of the Phd jury of Gaëtan Hadjeres (<i>Rapporteur</i>), Alexandre Bérard (<i>Head</i>), Olivier Ruas (<i>Rapporteur</i>), Valentina Zantedeschi.</p>
          </li>
          <li id="uid192">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> was <i>rapporteur</i> on the Phd jury of Elena Knyazeva, Université Paris-Saclay.</p>
          </li>
          <li id="uid193">
            <p noindent="true"><span class="smallcap" align="left">Mikaela Keller</span> was member of the recruitment committee for Assistant Professors
in Computer Science at Université of Lille and at Université de St-Étienne.</p>
          </li>
          <li id="uid194">
            <p noindent="true"><span class="smallcap" align="left">Mikaela Keller</span> was member of the Phd jury of Damien Fourure (Université de
St-Étienne) and of the HDR jury of Renaud Lopes (CHRU Lille).</p>
          </li>
          <li id="uid195">
            <p noindent="true"><span class="smallcap" align="left">Rémi Gilleron</span> was head of the PhD jury of Romain Warlop (Université de Lille).</p>
          </li>
          <li id="uid196">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> was a member of hiring committee for Junior Research
Scientist at Inria Lille.</p>
          </li>
          <li id="uid197">
            <p noindent="true"><span class="smallcap" align="left">Marc Tommasi</span> was member of the recruitment committee Assistant Professors in Computer Science at Université of Lille and for professor position at INSA de Lyon.</p>
          </li>
        </simplelist>
      </subsection>
    </subsection>
    <subsection id="uid198" level="1">
      <bodyTitle>Popularization</bodyTitle>
      <subsection id="uid199" level="2">
        <bodyTitle>Internal or external Inria responsibilities</bodyTitle>
        <simplelist>
          <li id="uid200">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> is the scientific mediation contact for Inria Lille center.</p>
          </li>
          <li id="uid201">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> served as committee member on the Inria Lille Commission
Emploi Recherche (CER).</p>
          </li>
          <li id="uid202">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> also served as committee member on Commission de Développement
Technologique (CDT).</p>
          </li>
          <li id="uid203">
            <p noindent="true"><span class="smallcap" align="left">Pascal Denis</span> is administrator of Inria membership to Linguistic Data
Consortium (LDC).</p>
          </li>
        </simplelist>
      </subsection>
      <subsection id="uid204" level="2">
        <bodyTitle>Articles and contents</bodyTitle>
        <simplelist>
          <li id="uid205">
            <p noindent="true"><span class="smallcap" align="left">Aurélien Bellet</span> and <span class="smallcap" align="left">Marc Tommasi</span> provided expertise for an upcoming TV program on Arte about new technologies.</p>
          </li>
        </simplelist>
      </subsection>
      <subsection id="uid206" level="2">
        <bodyTitle>Interventions</bodyTitle>
        <simplelist>
          <li id="uid207">
            <p noindent="true">National events: <span class="smallcap" align="left">Jan Ramon</span> and <span class="smallcap" align="left">Marc Tommasi</span> participate to a round-table meeting at the <i>Fête des libertés numériques</i> for the RGPD day <footnote id="uid208" id-text="7"><ref xlink:href="https://www.meshs.fr/page/donnees_personnelles_et_droits_et_libertes_numeriques" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>www.<allowbreak/>meshs.<allowbreak/>fr/<allowbreak/>page/<allowbreak/>donnees_personnelles_et_droits_et_libertes_numeriques</ref></footnote>.</p>
          </li>
          <li id="uid209">
            <p noindent="true">In educational institutions: <span class="smallcap" align="left">Marc Tommasi</span> gave a talk on privacy and machine learning in Journées polytech <footnote id="uid210" id-text="8"><ref xlink:href="http://www.polytech-lille.fr/big-data-machine-mearning-p11419.html#.WqlHjExFxPb" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>www.<allowbreak/>polytech-lille.<allowbreak/>fr/<allowbreak/>big-data-machine-mearning-p11419.<allowbreak/>html#.<allowbreak/>WqlHjExFxPb</ref></footnote>.</p>
          </li>
        </simplelist>
      </subsection>
    </subsection>
  </diffusion>
  <biblio id="bibliography" html="bibliography" numero="10" titre="Bibliography">
    
    <biblStruct id="magnet-2018-bid41" type="phdthesis" rend="year" n="cite:lefalher:tel-01824215">
      <identifiant type="hal" value="tel-01824215"/>
      <monogr>
        <title level="m">Characterizing Edges in Signed and Vector-Valued Graphs</title>
        <author>
          <persName>
            <foreName>Géraud</foreName>
            <surname>Le Falher</surname>
            <initial>G.</initial>
          </persName>
        </author>
        <imprint>
          <publisher>
            <orgName type="school">Université de Lille</orgName>
          </publisher>
          <dateStruct>
            <month>April</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/tel-01824215" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>tel-01824215</ref>
        </imprint>
      </monogr>
      <note type="typdoc">Theses</note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid28" type="article" rend="year" n="cite:cannoodt:hal-01814675">
      <identifiant type="doi" value="10.1371/ journal.pone.0195997"/>
      <identifiant type="hal" value="hal-01814675"/>
      <analytic>
        <title level="a">IncGraph: Incremental graphlet counting for topology optimisation</title>
        <author>
          <persName>
            <foreName>Robrecht</foreName>
            <surname>Cannoodt</surname>
            <initial>R.</initial>
          </persName>
          <persName>
            <foreName>Joeri</foreName>
            <surname>Ruyssinck</surname>
            <initial>J.</initial>
          </persName>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
          <persName>
            <foreName>Katleen</foreName>
            <surname>De Preter</surname>
            <initial>K.</initial>
          </persName>
          <persName>
            <foreName>Yvan</foreName>
            <surname>Saeys</surname>
            <initial>Y.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-editorial-board="yes" x-international-audience="yes" id="rid01553">
        <idno type="issn">1932-6203</idno>
        <title level="j">PLoS ONE</title>
        <imprint>
          <biblScope type="volume">13</biblScope>
          <biblScope type="number">4</biblScope>
          <dateStruct>
            <month>April</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01814675" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01814675</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid44" type="article" rend="year" n="cite:cappelle:hal-01959609">
      <identifiant type="hal" value="hal-01959609"/>
      <analytic>
        <title level="a">Facing the facts of fake: a distributional semantics and corpus annotation approach</title>
        <author>
          <persName key="magnet-2018-idp164272">
            <foreName>Bert</foreName>
            <surname>Cappelle</surname>
            <initial>B.</initial>
          </persName>
          <persName key="magnet-2018-idp156880">
            <foreName>Pascal</foreName>
            <surname>Denis</surname>
            <initial>P.</initial>
          </persName>
          <persName key="magnet-2018-idp166768">
            <foreName>Mikaela</foreName>
            <surname>Keller</surname>
            <initial>M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-editorial-board="yes" x-international-audience="yes" id="rid03418">
        <idno type="issn">2197-2796</idno>
        <title level="j">Yearbook of the German Cognitive Linguistics Association</title>
        <imprint>
          <dateStruct>
            <month>November</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.archives-ouvertes.fr/hal-01959609" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>archives-ouvertes.<allowbreak/>fr/<allowbreak/>hal-01959609</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid27" type="article" rend="year" n="cite:pelekis:hal-01814651">
      <identifiant type="hal" value="hal-01814651"/>
      <analytic>
        <title level="a">On the Bernstein-Hoeffding method</title>
        <author>
          <persName>
            <foreName>Christos</foreName>
            <surname>Pelekis</surname>
            <initial>C.</initial>
          </persName>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
          <persName>
            <foreName>Yuyi</foreName>
            <surname>Wang</surname>
            <initial>Y.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-editorial-board="yes" x-international-audience="yes" id="rid03417">
        <idno type="issn">0072-7466</idno>
        <title level="j">Bulletin of the Hellenic Mathematical Society</title>
        <imprint>
          <biblScope type="volume">62</biblScope>
          <dateStruct>
            <month>June</month>
            <year>2018</year>
          </dateStruct>
          <biblScope type="pages">31-43</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01814651" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01814651</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid29" type="article" rend="year" n="cite:ravkic:hal-01725971">
      <identifiant type="hal" value="hal-01725971"/>
      <analytic>
        <title level="a">Graph sampling with applications to estimating the number of pattern embeddings and the parameters of a statistical relational model</title>
        <author>
          <persName>
            <foreName>Irma</foreName>
            <surname>Ravkic</surname>
            <initial>I.</initial>
          </persName>
          <persName>
            <foreName>Martiň</foreName>
            <surname>Znidaršič</surname>
            <initial>M.</initial>
          </persName>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
          <persName>
            <foreName>Jesse</foreName>
            <surname>Davis</surname>
            <initial>J.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-editorial-board="yes" x-international-audience="yes" id="rid00457">
        <idno type="issn">1384-5810</idno>
        <title level="j">Data Mining and Knowledge Discovery</title>
        <imprint>
          <dateStruct>
            <year>2018</year>
          </dateStruct>
          <biblScope type="pages">36</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01725971" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01725971</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid30" type="article" rend="year" n="cite:schietgat:hal-01814669">
      <identifiant type="doi" value="10.1371/journal.pcbi.1006097"/>
      <identifiant type="hal" value="hal-01814669"/>
      <analytic>
        <title level="a">A machine learning based framework to identify and classify long terminal repeat retrotransposons</title>
        <author>
          <persName>
            <foreName>Leander</foreName>
            <surname>Schietgat</surname>
            <initial>L.</initial>
          </persName>
          <persName>
            <foreName>Celine</foreName>
            <surname>Vens</surname>
            <initial>C.</initial>
          </persName>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
          <persName>
            <foreName>Ricardo</foreName>
            <surname>Cerri</surname>
            <initial>R.</initial>
          </persName>
          <persName>
            <foreName>Carlos N</foreName>
            <surname>Fischer</surname>
            <initial>C. N.</initial>
          </persName>
          <persName>
            <foreName>Eduardo da</foreName>
            <surname>Costa</surname>
            <initial>E. d.</initial>
          </persName>
          <persName>
            <foreName>Claudia M A</foreName>
            <surname>Carareto</surname>
            <initial>C. M. A.</initial>
          </persName>
          <persName>
            <foreName>Hendrik</foreName>
            <surname>Blockeel</surname>
            <initial>H.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-editorial-board="yes" x-international-audience="yes" id="rid01552">
        <idno type="issn">1553-734X</idno>
        <title level="j">PLoS Computational Biology</title>
        <imprint>
          <biblScope type="volume">14</biblScope>
          <biblScope type="number">4</biblScope>
          <dateStruct>
            <month>April</month>
            <year>2018</year>
          </dateStruct>
          <biblScope type="pages">1-21</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01814669" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01814669</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid31" type="article" rend="year" n="cite:zheng:hal-01922994">
      <identifiant type="hal" value="hal-01922994"/>
      <analytic>
        <title level="a">A Distributed Frank-Wolfe Framework for Learning Low-Rank Matrices with the Trace Norm</title>
        <author>
          <persName>
            <foreName>Wenjie</foreName>
            <surname>Zheng</surname>
            <initial>W.</initial>
          </persName>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Patrick</foreName>
            <surname>Gallinari</surname>
            <initial>P.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-editorial-board="yes" x-international-audience="yes" id="rid01366">
        <idno type="issn">0885-6125</idno>
        <title level="j">Machine Learning</title>
        <imprint>
          <dateStruct>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01922994" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01922994</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid34" type="inproceedings" rend="year" n="cite:ailem:hal-01922985">
      <identifiant type="hal" value="hal-01922985"/>
      <analytic>
        <title level="a">A Probabilistic Model for Joint Learning of Word Embeddings from Texts and Images</title>
        <author>
          <persName key="magnet-2018-idp177136">
            <foreName>Melissa</foreName>
            <surname>Ailem</surname>
            <initial>M.</initial>
          </persName>
          <persName>
            <foreName>Bowen</foreName>
            <surname>Zhang</surname>
            <initial>B.</initial>
          </persName>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName key="magnet-2018-idp156880">
            <foreName>Pascal</foreName>
            <surname>Denis</surname>
            <initial>P.</initial>
          </persName>
          <persName>
            <foreName>Fei</foreName>
            <surname>Sha</surname>
            <initial>F.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">Conference on Empirical Methods in Natural Language Processing (EMNLP 2018)</title>
        <loc>Brussels, Belgium</loc>
        <imprint>
          <dateStruct>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01922985" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01922985</ref>
        </imprint>
        <meeting id="cid47305">
          <title>Conference on Empirical Methods in Natural Language Processing</title>
          <num>2018</num>
          <abbr type="sigle">EMNLP</abbr>
        </meeting>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid32" type="inproceedings" rend="year" n="cite:bellet:hal-01745796">
      <identifiant type="hal" value="hal-01745796"/>
      <analytic>
        <title level="a">Personalized and Private Peer-to-Peer Machine Learning</title>
        <author>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Rachid</foreName>
            <surname>Guerraoui</surname>
            <initial>R.</initial>
          </persName>
          <persName>
            <foreName>Mahsa</foreName>
            <surname>Taziki</surname>
            <initial>M.</initial>
          </persName>
          <persName key="magnet-2018-idp169248">
            <foreName>Marc</foreName>
            <surname>Tommasi</surname>
            <initial>M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">AISTATS 2018 - 21st International Conference on Artificial Intelligence and Statistics</title>
        <loc>Lanzarote, Spain</loc>
        <imprint>
          <dateStruct>
            <month>April</month>
            <year>2018</year>
          </dateStruct>
          <biblScope type="pages">1-20</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01745796" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01745796</ref>
        </imprint>
        <meeting id="cid388734">
          <title>International Conference on Artificial Intelligence and Statistics</title>
          <num>21</num>
          <abbr type="sigle">AISTATS</abbr>
        </meeting>
      </monogr>
      <note type="bnote">
        <ref xlink:href="https://arxiv.org/abs/1705.08435" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>arxiv.<allowbreak/>org/<allowbreak/>abs/<allowbreak/>1705.<allowbreak/>08435</ref>
      </note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid40" type="inproceedings" rend="year" n="cite:cesabianchi:hal-01916981">
      <identifiant type="hal" value="hal-01916981"/>
      <analytic>
        <title level="a">Nonstochastic Bandits with Composite Anonymous Feedback</title>
        <author>
          <persName>
            <foreName>Nicolo</foreName>
            <surname>Cesa-Bianchi</surname>
            <initial>N.</initial>
          </persName>
          <persName key="magnet-2018-idp159344">
            <foreName>Claudio</foreName>
            <surname>Gentile</surname>
            <initial>C.</initial>
          </persName>
          <persName>
            <foreName>Yishay</foreName>
            <surname>Mansour</surname>
            <initial>Y.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">COLT 2018 - 31st Annual Conference on Learning Theory</title>
        <loc>Stockholm, Sweden</loc>
        <imprint>
          <biblScope type="volume">75</biblScope>
          <dateStruct>
            <month>July</month>
            <year>2018</year>
          </dateStruct>
          <biblScope type="pages">1 - 23</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01916981" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01916981</ref>
        </imprint>
        <meeting id="cid29437">
          <title>Annual Conference on Learning Theory</title>
          <num>31</num>
          <abbr type="sigle">COLT</abbr>
        </meeting>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid35" type="inproceedings" rend="year" n="cite:dehouck:hal-01943934">
      <identifiant type="hal" value="hal-01943934"/>
      <analytic>
        <title level="a">A Framework for Understanding the Role of Morphology in Universal Dependency Parsing</title>
        <author>
          <persName key="magnet-2018-idp186944">
            <foreName>Mathieu</foreName>
            <surname>Dehouck</surname>
            <initial>M.</initial>
          </persName>
          <persName key="magnet-2018-idp156880">
            <foreName>Pascal</foreName>
            <surname>Denis</surname>
            <initial>P.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">EMNLP 2018 - Conference on Empirical Methods in Natural Language Processing</title>
        <loc>Brussels, Belgium</loc>
        <title level="s">Proceedings of EMNLP 2018</title>
        <imprint>
          <dateStruct>
            <month>October</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.archives-ouvertes.fr/hal-01943934" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>archives-ouvertes.<allowbreak/>fr/<allowbreak/>hal-01943934</ref>
        </imprint>
        <meeting id="cid47305">
          <title>Conference on Empirical Methods in Natural Language Processing</title>
          <num>2018</num>
          <abbr type="sigle">EMNLP</abbr>
        </meeting>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid37" type="inproceedings" rend="year" n="cite:pasteris:hal-01916976">
      <identifiant type="hal" value="hal-01916976"/>
      <analytic>
        <title level="a">On Similarity Prediction and Pairwise Clustering</title>
        <author>
          <persName>
            <foreName>Stephen</foreName>
            <surname>Pasteris</surname>
            <initial>S.</initial>
          </persName>
          <persName key="magnet-2018-idp174656">
            <foreName>Fabio</foreName>
            <surname>Vitale</surname>
            <initial>F.</initial>
          </persName>
          <persName key="magnet-2018-idp159344">
            <foreName>Claudio</foreName>
            <surname>Gentile</surname>
            <initial>C.</initial>
          </persName>
          <persName>
            <foreName>Mark</foreName>
            <surname>Herbster</surname>
            <initial>M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">ALT 2018 - 29th International Conference on Algorithmic Learning Theory</title>
        <loc>Lanzarote, Spain</loc>
        <imprint>
          <biblScope type="volume">83</biblScope>
          <dateStruct>
            <month>April</month>
            <year>2018</year>
          </dateStruct>
          <biblScope type="pages">1 - 28</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01916976" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01916976</ref>
        </imprint>
        <meeting id="cid110465">
          <title>International Conference on Algorithmic Learning Theory</title>
          <num>29</num>
          <abbr type="sigle">ALT</abbr>
        </meeting>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid36" type="inproceedings" rend="year" n="cite:vitale:hal-01916979">
      <identifiant type="hal" value="hal-01916979"/>
      <analytic>
        <title level="a">Online Reciprocal Recommendation with Theoretical Performance Guarantees</title>
        <author>
          <persName key="magnet-2018-idp174656">
            <foreName>Fabio</foreName>
            <surname>Vitale</surname>
            <initial>F.</initial>
          </persName>
          <persName>
            <foreName>Nikos</foreName>
            <surname>Parotsidis</surname>
            <initial>N.</initial>
          </persName>
          <persName key="magnet-2018-idp159344">
            <foreName>Claudio</foreName>
            <surname>Gentile</surname>
            <initial>C.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">NIPS 2018 - 32nd Conference on Neural Information Processing Systems</title>
        <loc>Montreal, Canada</loc>
        <imprint>
          <dateStruct>
            <month>December</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01916979" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01916979</ref>
        </imprint>
        <meeting id="cid29560">
          <title>Annual Conference on Neural Information Processing Systems</title>
          <num>32</num>
          <abbr type="sigle">NIPS</abbr>
        </meeting>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid38" type="inproceedings" rend="year" n="cite:vogel:hal-01922988">
      <identifiant type="hal" value="hal-01922988"/>
      <analytic>
        <title level="a">A Probabilistic Theory of Supervised Similarity Learning for Pointwise ROC Curve Optimization</title>
        <author>
          <persName>
            <foreName>Robin</foreName>
            <surname>Vogel</surname>
            <initial>R.</initial>
          </persName>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Stéphan</foreName>
            <surname>Clémençon</surname>
            <initial>S.</initial>
          </persName>
        </author>
      </analytic>
      <monogr x-scientific-popularization="no" x-international-audience="yes" x-proceedings="yes" x-invited-conference="no" x-editorial-board="yes">
        <title level="m">International Conference on Machine Learning (ICML 2018)</title>
        <loc>Stockholm, Sweden</loc>
        <imprint>
          <dateStruct>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01922988" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01922988</ref>
        </imprint>
        <meeting id="cid32516">
          <title>International Conference on Machine Learning</title>
          <num>35</num>
          <abbr type="sigle">ICML</abbr>
        </meeting>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid33" type="techreport" rend="year" n="cite:dellenbach:hal-01923000">
      <identifiant type="hal" value="hal-01923000"/>
      <monogr>
        <title level="m">Hiding in the Crowd: A Massively Distributed Algorithm for Private Averaging with Malicious Adversaries</title>
        <author>
          <persName>
            <foreName>Pierre</foreName>
            <surname>Dellenbach</surname>
            <initial>P.</initial>
          </persName>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
        </author>
        <imprint>
          <publisher>
            <orgName type="institution">Inria</orgName>
          </publisher>
          <dateStruct>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01923000" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01923000</ref>
        </imprint>
      </monogr>
      <note type="typdoc">Research Report</note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid39" type="techreport" rend="year" n="cite:liu:hal-01923006">
      <identifiant type="hal" value="hal-01923006"/>
      <monogr>
        <title level="m">Escaping the Curse of Dimensionality in Similarity Learning: Efficient Frank-Wolfe Algorithm and Generalization Bounds</title>
        <author>
          <persName>
            <foreName>Kuan</foreName>
            <surname>Liu</surname>
            <initial>K.</initial>
          </persName>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
        </author>
        <imprint>
          <publisher>
            <orgName type="institution">Inria</orgName>
          </publisher>
          <dateStruct>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01923006" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01923006</ref>
        </imprint>
      </monogr>
      <note type="typdoc">Research Report</note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid45" type="unpublished" rend="year" n="cite:capelli:hal-01981553">
      <identifiant type="hal" value="hal-01981553"/>
      <monogr>
        <title level="m">Dependency Weighted Aggregation on Factorized Databases</title>
        <author>
          <persName key="links-2018-idp164416">
            <foreName>Florent</foreName>
            <surname>Capelli</surname>
            <initial>F.</initial>
          </persName>
          <persName key="links-2018-idp186016">
            <foreName>Nicolas</foreName>
            <surname>Crosetti</surname>
            <initial>N.</initial>
          </persName>
          <persName key="links-2018-idp158976">
            <foreName>Joachim</foreName>
            <surname>Niehren</surname>
            <initial>J.</initial>
          </persName>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
        </author>
        <imprint>
          <dateStruct>
            <month>January</month>
            <year>2019</year>
          </dateStruct>
          <ref xlink:href="https://hal.archives-ouvertes.fr/hal-01981553" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>archives-ouvertes.<allowbreak/>fr/<allowbreak/>hal-01981553</ref>
        </imprint>
      </monogr>
      <note type="bnote"><ref xlink:href="https://arxiv.org/abs/1901.03633" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>arxiv.<allowbreak/>org/<allowbreak/>abs/<allowbreak/>1901.<allowbreak/>03633</ref> - working paper or preprint</note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid43" type="unpublished" rend="year" n="cite:maillard:hal-01945034">
      <identifiant type="hal" value="hal-01945034"/>
      <monogr>
        <title level="m">Upper Confidence Reinforcement Learning exploiting state-action equivalence</title>
        <author>
          <persName key="sequel-2018-idp161536">
            <foreName>Odalric-Ambrym</foreName>
            <surname>Maillard</surname>
            <initial>O.-A.</initial>
          </persName>
          <persName key="magnet-2018-idp184512">
            <foreName>Mahsa</foreName>
            <surname>Asadi</surname>
            <initial>M.</initial>
          </persName>
        </author>
        <imprint>
          <dateStruct>
            <month>December</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.archives-ouvertes.fr/hal-01945034" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>archives-ouvertes.<allowbreak/>fr/<allowbreak/>hal-01945034</ref>
        </imprint>
      </monogr>
      <note type="bnote">working paper or preprint</note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid42" type="misc" rend="year" n="cite:ramon:hal-01879941">
      <identifiant type="hal" value="hal-01879941"/>
      <monogr x-scientific-popularization="no">
        <title level="m">Exploiting traffic data</title>
        <author>
          <persName key="magnet-2018-idp161824">
            <foreName>Jan</foreName>
            <surname>Ramon</surname>
            <initial>J.</initial>
          </persName>
        </author>
        <imprint>
          <dateStruct>
            <month>December</month>
            <year>2018</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01879941" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01879941</ref>
        </imprint>
      </monogr>
      <note type="bnote">Presentation given at Inria Meetup (Lille)</note>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid2" type="inproceedings" rend="foot" n="footcite:AlexandrescuKirchhof2007">
      <analytic>
        <title level="a">Graph-based learning for phonetic classification</title>
        <author>
          <persName>
            <foreName>Andrei</foreName>
            <surname>Alexandrescu</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Katrin</foreName>
            <surname>Kirchhoff</surname>
            <initial>K.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">IEEE Workshop on Automatic Speech Recognition &amp; Understanding, ASRU 2007, Kyoto, Japan, December 9-13, 2007</title>
        <imprint>
          <dateStruct>
            <year>2007</year>
          </dateStruct>
          <biblScope type="pages">359-364</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid16" type="inproceedings" rend="foot" n="footcite:BalcanEtAl2005">
      <analytic>
        <title level="a">Person Identification in Webcam Images: An Application of Semi-Supervised Learning</title>
        <author>
          <persName>
            <foreName>Maria-Florina</foreName>
            <surname>Balcan</surname>
            <initial>M.-F.</initial>
          </persName>
          <persName>
            <foreName>Avrim</foreName>
            <surname>Blum</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Patrick Pakyan</foreName>
            <surname>Choi</surname>
            <initial>P. P.</initial>
          </persName>
          <persName>
            <foreName>John</foreName>
            <surname>Lafferty</surname>
            <initial>J.</initial>
          </persName>
          <persName>
            <foreName>Brian</foreName>
            <surname>Pantano</surname>
            <initial>B.</initial>
          </persName>
          <persName>
            <foreName>Mugizi R.</foreName>
            <surname>Rwebangira</surname>
            <initial>M. R.</initial>
          </persName>
          <persName>
            <foreName>Xiaojin</foreName>
            <surname>Zhu</surname>
            <initial>X.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">ICML2005 Workshop on Learning with Partially Classified Training Data</title>
        <imprint>
          <dateStruct>
            <year>2005</year>
          </dateStruct>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid17" type="article" rend="foot" n="footcite:BelkinNiyogi2008">
      <analytic>
        <title level="a">Towards a Theoretical Foundation for Laplacian-Based Manifold Methods</title>
        <author>
          <persName>
            <foreName>Mikhail</foreName>
            <surname>Belkin</surname>
            <initial>M.</initial>
          </persName>
          <persName>
            <foreName>Partha</foreName>
            <surname>Niyogi</surname>
            <initial>P.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="j">Journal of Computer and System Sciences</title>
        <imprint>
          <biblScope type="volume">74</biblScope>
          <biblScope type="number">8</biblScope>
          <dateStruct>
            <year>2008</year>
          </dateStruct>
          <biblScope type="pages">1289-1308</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid18" type="article" rend="foot" n="footcite:BelletHabrardSebban2013">
      <analytic>
        <title level="a">A Survey on Metric Learning for Feature Vectors and Structured Data</title>
        <author>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Amaury</foreName>
            <surname>Habrard</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Marc</foreName>
            <surname>Sebban</surname>
            <initial>M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="j">CoRR</title>
        <imprint>
          <biblScope type="volume">abs/1306.6709</biblScope>
          <dateStruct>
            <year>2013</year>
          </dateStruct>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid19" type="book" rend="foot" n="footcite:Bellet2015c">
      <monogr>
        <title level="m">Metric Learning</title>
        <author>
          <persName key="magnet-2018-idp154416">
            <foreName>Aurélien</foreName>
            <surname>Bellet</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Amaury</foreName>
            <surname>Habrard</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Marc</foreName>
            <surname>Sebban</surname>
            <initial>M.</initial>
          </persName>
        </author>
        <imprint>
          <publisher>
            <orgName>Morgan &amp; Claypool Publishers</orgName>
          </publisher>
          <dateStruct>
            <year>2015</year>
          </dateStruct>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid23" type="article" rend="foot" n="footcite:BickelAndChen09">
      <analytic>
        <title level="a">A nonparametric view of network models and Newman–Girvan and other modularities</title>
        <author>
          <persName>
            <foreName>Peter J.</foreName>
            <surname>Bickel</surname>
            <initial>P. J.</initial>
          </persName>
          <persName>
            <foreName>Aiyou</foreName>
            <surname>Chen</surname>
            <initial>A.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="j">Proceedings of the National Academy of Sciences</title>
        <imprint>
          <biblScope type="volume">106</biblScope>
          <dateStruct>
            <year>2009</year>
          </dateStruct>
          <biblScope type="pages">21068–21073</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid24" type="book" rend="foot" n="footcite:blau1977inequality">
      <monogr>
        <title level="m">Inequality and Heterogeneity: A Primitive Theory of Social Structure</title>
        <author>
          <persName>
            <foreName>P.M.</foreName>
            <surname>Blau</surname>
            <initial>P.</initial>
          </persName>
        </author>
        <imprint>
          <publisher>
            <orgName>MACMILLAN Company</orgName>
          </publisher>
          <dateStruct>
            <year>1977</year>
          </dateStruct>
          <ref xlink:href="http://books.google.fr/books?id=jvq2AAAAIAAJ" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>books.<allowbreak/>google.<allowbreak/>fr/<allowbreak/>books?id=jvq2AAAAIAAJ</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid9" type="inproceedings" rend="foot" n="footcite:braud:hal-01017151">
      <identifiant type="hal" value="hal-01017151"/>
      <analytic>
        <title level="a">Combining Natural and Artificial Examples to Improve Implicit Discourse Relation Identification</title>
        <author>
          <persName>
            <foreName>Chloé</foreName>
            <surname>Braud</surname>
            <initial>C.</initial>
          </persName>
          <persName key="magnet-2018-idp156880">
            <foreName>Pascal</foreName>
            <surname>Denis</surname>
            <initial>P.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">coling</title>
        <loc>Dublin, Ireland</loc>
        <imprint>
          <dateStruct>
            <month>August</month>
            <year>2014</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01017151" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01017151</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid13" type="inproceedings" rend="foot" n="footcite:ChangYeung2006">
      <identifiant type="doi" value="10.1109/CVPR.2006.128"/>
      <analytic>
        <title level="a">Graph Laplacian Kernels for Object Classification from a Single Example</title>
        <author>
          <persName>
            <foreName>Hong</foreName>
            <surname>Chang</surname>
            <initial>H.</initial>
          </persName>
          <persName>
            <foreName>Dit-Yan</foreName>
            <surname>Yeung</surname>
            <initial>D.-Y.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Volume 2</title>
        <loc>Washington, DC, USA</loc>
        <title level="s">CVPR '06</title>
        <imprint>
          <publisher>
            <orgName>IEEE Computer Society</orgName>
          </publisher>
          <dateStruct>
            <year>2006</year>
          </dateStruct>
          <biblScope type="pages">2011–2016</biblScope>
          <ref xlink:href="http://dx.doi.org/10.1109/CVPR.2006.128" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>dx.<allowbreak/>doi.<allowbreak/>org/<allowbreak/>10.<allowbreak/>1109/<allowbreak/>CVPR.<allowbreak/>2006.<allowbreak/>128</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid10" type="inproceedings" rend="foot" n="footcite:chatel:hal-01017269">
      <identifiant type="doi" value="10.1007/978-3-662-44848-9_16"/>
      <identifiant type="hal" value="hal-01017269"/>
      <analytic>
        <title level="a">Fast Gaussian Pairwise Constrained Spectral Clustering</title>
        <author>
          <persName>
            <foreName>David</foreName>
            <surname>Chatel</surname>
            <initial>D.</initial>
          </persName>
          <persName key="magnet-2018-idp156880">
            <foreName>Pascal</foreName>
            <surname>Denis</surname>
            <initial>P.</initial>
          </persName>
          <persName key="magnet-2018-idp169248">
            <foreName>Marc</foreName>
            <surname>Tommasi</surname>
            <initial>M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">ECML/PKDD - 7th European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
        <loc>Nancy, France</loc>
        <imprint>
          <dateStruct>
            <month>September</month>
            <year>2014</year>
          </dateStruct>
          <biblScope type="pages">242 - 257</biblScope>
          <ref xlink:href="https://hal.inria.fr/hal-01017269" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01017269</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid6" type="inproceedings" rend="foot" n="footcite:DasPetrov2011">
      <analytic>
        <title level="a">Unsupervised Part-of-Speech Tagging with Bilingual Graph-Based Projections</title>
        <author>
          <persName>
            <foreName>Dipanjan</foreName>
            <surname>Das</surname>
            <initial>D.</initial>
          </persName>
          <persName>
            <foreName>Slav</foreName>
            <surname>Petrov</surname>
            <initial>S.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">ACL</title>
        <imprint>
          <dateStruct>
            <year>2011</year>
          </dateStruct>
          <biblScope type="pages">600-609</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid8" type="inproceedings" rend="foot" n="footcite:denis:muller:2011">
      <identifiant type="hal" value="inria-00614765"/>
      <analytic>
        <title level="a">Predicting globally-coherent temporal structures from texts via endpoint inference and graph decomposition</title>
        <author>
          <persName key="magnet-2018-idp156880">
            <foreName>Pascal</foreName>
            <surname>Denis</surname>
            <initial>P.</initial>
          </persName>
          <persName>
            <foreName>Philippe</foreName>
            <surname>Muller</surname>
            <initial>P.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">IJCAI-11 - International Joint Conference on Artificial Intelligence</title>
        <loc>Barcelone, Espagne</loc>
        <imprint>
          <dateStruct>
            <year>2011</year>
          </dateStruct>
          <ref xlink:href="http://hal.inria.fr/inria-00614765" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>inria-00614765</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid7" type="inproceedings" rend="foot" n="footcite:FernandesBrefeld2011">
      <analytic>
        <title level="a">Learning from Partially Annotated Sequences</title>
        <author>
          <persName>
            <foreName>Eraldo R.</foreName>
            <surname>Fernandes</surname>
            <initial>E. R.</initial>
          </persName>
          <persName>
            <foreName>Ulf</foreName>
            <surname>Brefeld</surname>
            <initial>U.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">ECML/PKDD</title>
        <imprint>
          <dateStruct>
            <year>2011</year>
          </dateStruct>
          <biblScope type="pages">407-422</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid14" type="inproceedings" rend="foot" n="footcite:GoldbergZhu2006">
      <analytic>
        <title level="a">Seeing stars when there aren't many stars: graph-based semi-supervised learning for sentiment categorization</title>
        <author>
          <persName>
            <foreName>Andrew B.</foreName>
            <surname>Goldberg</surname>
            <initial>A. B.</initial>
          </persName>
          <persName>
            <foreName>Xiaojin</foreName>
            <surname>Zhu</surname>
            <initial>X.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proceedings of the First Workshop on Graph Based Methods for Natural Language Processing</title>
        <loc>Stroudsburg, PA, USA</loc>
        <title level="s">TextGraphs-1</title>
        <imprint>
          <publisher>
            <orgName>Association for Computational Linguistics</orgName>
          </publisher>
          <dateStruct>
            <year>2006</year>
          </dateStruct>
          <biblScope type="pages">45–52</biblScope>
          <ref xlink:href="http://dl.acm.org/citation.cfm?id=1654758.1654769" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>dl.<allowbreak/>acm.<allowbreak/>org/<allowbreak/>citation.<allowbreak/>cfm?id=1654758.<allowbreak/>1654769</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid22" type="book" rend="foot" n="footcite:Goldenberg2010survey">
      <monogr>
        <title level="m">A Survey of Statistical Network Models</title>
        <title level="s">Foundations and trends in machine learning</title>
        <author>
          <persName>
            <foreName>A</foreName>
            <surname>Goldenberg</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>A X</foreName>
            <surname>Zheng</surname>
            <initial>A. X.</initial>
          </persName>
          <persName>
            <foreName>S E</foreName>
            <surname>Fienberg</surname>
            <initial>S. E.</initial>
          </persName>
        </author>
        <imprint>
          <publisher>
            <orgName>Now Publishers</orgName>
          </publisher>
          <dateStruct>
            <year>2010</year>
          </dateStruct>
          <ref xlink:href="http://books.google.fr/books?id=gPGgcOf95moC" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>books.<allowbreak/>google.<allowbreak/>fr/<allowbreak/>books?id=gPGgcOf95moC</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid20" type="inproceedings" rend="foot" n="footcite:GomezLeskovecKrause2010">
      <analytic>
        <title level="a">Inferring networks of diffusion and influence</title>
        <author>
          <persName>
            <foreName>Manuel</foreName>
            <surname>Gomez-Rodriguez</surname>
            <initial>M.</initial>
          </persName>
          <persName>
            <foreName>Jure</foreName>
            <surname>Leskovec</surname>
            <initial>J.</initial>
          </persName>
          <persName>
            <foreName>Andreas</foreName>
            <surname>Krause</surname>
            <initial>A.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proc. of KDD</title>
        <imprint>
          <dateStruct>
            <year>2010</year>
          </dateStruct>
          <biblScope type="pages">1019-1028</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid25" type="article" rend="foot" n="footcite:mcpherson2001">
      <identifiant type="doi" value="10.1146/annurev.soc.27.1.415"/>
      <analytic>
        <title level="a">Birds of a Feather: Homophily in Social Networks</title>
        <author>
          <persName>
            <foreName>Miller</foreName>
            <surname>McPherson</surname>
            <initial>M.</initial>
          </persName>
          <persName>
            <foreName>Lynn S</foreName>
            <surname>Lovin</surname>
            <initial>L. S.</initial>
          </persName>
          <persName>
            <foreName>James M</foreName>
            <surname>Cook</surname>
            <initial>J. M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="j">Annual Review of Sociology</title>
        <imprint>
          <biblScope type="volume">27</biblScope>
          <biblScope type="number">1</biblScope>
          <dateStruct>
            <year>2001</year>
          </dateStruct>
          <biblScope type="pages">415–444</biblScope>
          <ref xlink:href="http://dx.doi.org/10.1146/annurev.soc.27.1.415" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>dx.<allowbreak/>doi.<allowbreak/>org/<allowbreak/>10.<allowbreak/>1146/<allowbreak/>annurev.<allowbreak/>soc.<allowbreak/>27.<allowbreak/>1.<allowbreak/>415</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid0" type="incollection" rend="foot" n="footcite:NenkovaMcKeown2012">
      <analytic>
        <title level="a">A Survey of Text Summarization Techniques</title>
        <author>
          <persName>
            <foreName>Ani</foreName>
            <surname>Nenkova</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Kathleen</foreName>
            <surname>McKeown</surname>
            <initial>K.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Mining Text Data</title>
        <imprint>
          <publisher>
            <orgName>Springer</orgName>
          </publisher>
          <dateStruct>
            <year>2012</year>
          </dateStruct>
          <biblScope type="pages">43-76</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid26" type="inproceedings" rend="foot" n="footcite:ricatte:hal-01017025">
      <identifiant type="hal" value="hal-01017025"/>
      <analytic>
        <title level="a">Hypernode Graphs for Spectral Learning on Binary Relations over Sets</title>
        <author>
          <persName>
            <foreName>Thomas</foreName>
            <surname>Ricatte</surname>
            <initial>T.</initial>
          </persName>
          <persName key="magnet-2018-idp223968">
            <foreName>Rémi</foreName>
            <surname>Gilleron</surname>
            <initial>R.</initial>
          </persName>
          <persName key="magnet-2018-idp169248">
            <foreName>Marc</foreName>
            <surname>Tommasi</surname>
            <initial>M.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">ECML/PKDD - 7th European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
        <loc>Nancy, France</loc>
        <title level="s">Machine Learning and Knowledge Discovery in Databases</title>
        <imprint>
          <dateStruct>
            <month>September</month>
            <year>2014</year>
          </dateStruct>
          <ref xlink:href="https://hal.inria.fr/hal-01017025" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">https://<allowbreak/>hal.<allowbreak/>inria.<allowbreak/>fr/<allowbreak/>hal-01017025</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid12" type="article" rend="foot" n="footcite:ShinTsudaScholkopf2009">
      <identifiant type="doi" value="10.1016/j.eswa.2008.01.006"/>
      <analytic>
        <title level="a">Protein functional class prediction with a combined graph</title>
        <author>
          <persName>
            <foreName>Hyunjung</foreName>
            <surname>Shin</surname>
            <initial>H.</initial>
          </persName>
          <persName>
            <foreName>Koji</foreName>
            <surname>Tsuda</surname>
            <initial>K.</initial>
          </persName>
          <persName>
            <foreName>Bernhard</foreName>
            <surname>Schölkopf</surname>
            <initial>B.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="j">Expert Syst. Appl.</title>
        <imprint>
          <biblScope type="volume">36</biblScope>
          <biblScope type="number">2</biblScope>
          <dateStruct>
            <month>March</month>
            <year>2009</year>
          </dateStruct>
          <biblScope type="pages">3284–3292</biblScope>
          <ref xlink:href="http://dx.doi.org/10.1016/j.eswa.2008.01.006" location="extern" xlink:type="simple" xlink:show="replace" xlink:actuate="onRequest">http://<allowbreak/>dx.<allowbreak/>doi.<allowbreak/>org/<allowbreak/>10.<allowbreak/>1016/<allowbreak/>j.<allowbreak/>eswa.<allowbreak/>2008.<allowbreak/>01.<allowbreak/>006</ref>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid1" type="inproceedings" rend="foot" n="footcite:SinghEtAl2011">
      <analytic>
        <title level="a">Large-Scale Cross-Document Coreference Using Distributed Inference and Hierarchical Models</title>
        <author>
          <persName>
            <foreName>Sameer</foreName>
            <surname>Singh</surname>
            <initial>S.</initial>
          </persName>
          <persName>
            <foreName>Amarnag</foreName>
            <surname>Subramanya</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Fernando C. N.</foreName>
            <surname>Pereira</surname>
            <initial>F. C. N.</initial>
          </persName>
          <persName>
            <foreName>Andrew</foreName>
            <surname>McCallum</surname>
            <initial>A.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">ACL</title>
        <imprint>
          <dateStruct>
            <year>2011</year>
          </dateStruct>
          <biblScope type="pages">793-803</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid4" type="inproceedings" rend="foot" n="footcite:SperiosuEtAl2011">
      <analytic>
        <title level="a">Twitter Polarity Classification with Label Propagation over Lexical Links and the Follower Graph</title>
        <author>
          <persName>
            <foreName>Michael</foreName>
            <surname>Speriosu</surname>
            <initial>M.</initial>
          </persName>
          <persName>
            <foreName>Nikita</foreName>
            <surname>Sudan</surname>
            <initial>N.</initial>
          </persName>
          <persName>
            <foreName>Sid</foreName>
            <surname>Upadhyay</surname>
            <initial>S.</initial>
          </persName>
          <persName>
            <foreName>Jason</foreName>
            <surname>Baldridge</surname>
            <initial>J.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proceedings of the First Workshop on Unsupervised Methods in NLP</title>
        <loc>Edinburgh, Scotland</loc>
        <imprint>
          <dateStruct>
            <year>2011</year>
          </dateStruct>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid3" type="inproceedings" rend="foot" n="footcite:SubramanyaEtAl2010">
      <analytic>
        <title level="a">Efficient Graph-Based Semi-Supervised Learning of Structured Tagging Models</title>
        <author>
          <persName>
            <foreName>Amarnag</foreName>
            <surname>Subramanya</surname>
            <initial>A.</initial>
          </persName>
          <persName>
            <foreName>Slav</foreName>
            <surname>Petrov</surname>
            <initial>S.</initial>
          </persName>
          <persName>
            <foreName>Fernando C. N.</foreName>
            <surname>Pereira</surname>
            <initial>F. C. N.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">EMNLP</title>
        <imprint>
          <dateStruct>
            <year>2010</year>
          </dateStruct>
          <biblScope type="pages">167-176</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid21" type="inproceedings" rend="foot" n="footcite:VitaleCesa-BianchiGentileZapellaNIPS2011">
      <analytic>
        <title level="a">See the Tree Through the Lines: The Shazoo Algorithm</title>
        <author>
          <persName key="magnet-2018-idp174656">
            <foreName>Fabio</foreName>
            <surname>Vitale</surname>
            <initial>F.</initial>
          </persName>
          <persName>
            <foreName>Nicolò</foreName>
            <surname>Cesa-Bianchi</surname>
            <initial>N.</initial>
          </persName>
          <persName key="magnet-2018-idp159344">
            <foreName>Claudio</foreName>
            <surname>Gentile</surname>
            <initial>C.</initial>
          </persName>
          <persName>
            <foreName>Giovanni</foreName>
            <surname>Zappella</surname>
            <initial>G.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proc of NIPS</title>
        <imprint>
          <dateStruct>
            <year>2011</year>
          </dateStruct>
          <biblScope type="pages">1584-1592</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid11" type="inproceedings" rend="foot" n="footcite:WangEtAl2011">
      <analytic>
        <title level="a">The Utility of Discourse Structure in Identifying Resolved Threads in Technical User Forums</title>
        <author>
          <persName key="tea-2018-idp151376">
            <foreName>Li</foreName>
            <surname>Wang</surname>
            <initial>L.</initial>
          </persName>
          <persName>
            <foreName>Su Nam</foreName>
            <surname>Kim</surname>
            <initial>S. N.</initial>
          </persName>
          <persName>
            <foreName>Timothy</foreName>
            <surname>Baldwin</surname>
            <initial>T.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">COLING</title>
        <imprint>
          <dateStruct>
            <year>2012</year>
          </dateStruct>
          <biblScope type="pages">2739-2756</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid5" type="inproceedings" rend="foot" n="footcite:LiuKirchhoff2013">
      <analytic>
        <title level="a">Graph-Based Semi-Supervised Learning for Phone and Segment Classification</title>
        <author>
          <persName>
            <foreName>Katrin Kirchhoff</foreName>
            <surname>Yuzong Liu</surname>
            <initial>K. K.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proceedings of Interspeech</title>
        <loc>Lyon, France</loc>
        <imprint>
          <dateStruct>
            <year>2013</year>
          </dateStruct>
        </imprint>
      </monogr>
    </biblStruct>
    
    <biblStruct id="magnet-2018-bid15" type="inproceedings" rend="foot" n="footcite:ZhuEtAl2003">
      <analytic>
        <title level="a">Semi-Supervised Learning Using Gaussian Fields and Harmonic Functions</title>
        <author>
          <persName>
            <foreName>Xiaojin</foreName>
            <surname>Zhu</surname>
            <initial>X.</initial>
          </persName>
          <persName>
            <foreName>Zoubin</foreName>
            <surname>Ghahramani</surname>
            <initial>Z.</initial>
          </persName>
          <persName>
            <foreName>John</foreName>
            <surname>Lafferty</surname>
            <initial>J.</initial>
          </persName>
        </author>
      </analytic>
      <monogr>
        <title level="m">Proc. of ICML</title>
        <imprint>
          <dateStruct>
            <year>2003</year>
          </dateStruct>
          <biblScope type="pages">912-919</biblScope>
        </imprint>
      </monogr>
    </biblStruct>
  </biblio>
</raweb>
