Ghent University – iMinds – Multimedia Lab

Ghent University – iMinds – Multimedia Lab
Master thesis subjects 2015 - 2016
Knowledge on Webscale
Assessing Open Source Code Trustworthiness through Version History
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Tom De Nies, Ruben Verborgh, Miel Vander Sande
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: tom.denies@ugent.be
Keywords: Trust, Version Control Systems, Git, Semantic Web
Location: home, Zuiderpoort
Problem definition:
An increasing amount of code is being shared on the Web, thanks to various open source initiatives
such as npm, sourceforge, Google Code, etc. However, with all this code being pushed to the Web,
and the possibility for anyone to contribute to any project, the quality of the resulting software is not
always optimal. Furthermore, manually judging whether or not to trust a piece of code is a timeconsuming and non-exact process, while programmers should be focusing on writing contributions
of their own. Therefore, there is a need for an automatic method to help programmers decide
whether or not they can trust a certain piece of code.
Goals:
To achieve this, the student will exploit the information contained within a Version Control System
(VCS), such as Git. More specifically, the ‘provenance’ (also referred to as ‘data lineage’) of the
code will be exposed using a tool such as Git2PROV. Then, the student will create a method to
automatically reason of this provenance, to give the end-user an assessment of the trustworthiness
of each version of the code. An extensive literature review will need to be conducted, to find out
which criteria influence trustworthiness, and how they can be inferred. Finally, the thesis should
result in a lightweight, user-friendly demonstrator that can easily be used by programmers in their
daily workflow.
Towards a Trusted Web by Tracing the Origins of Composite Web Pages
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Tom De Nies and Ruben Verborgh
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: tom.denies@ugent.be
Keywords: Trust, Version Control Systems, Git, Semantic Web
Location: home, Zuiderpoort
Problem definition:
While openness is one of the core foundations of the Web, it has caused such an abundance of
heterogeneous content that it becomes unclear for humans if they can trust the content they see on
web pages. Furthermore, web pages are often littered with tracking mechanisms, such as hidden
pixels, cookies, etc. The first step in deciding whether or not to trust a web page is finding out where
its contents come from, who made/edited them and whether these sources are to be trusted. This is
what’s known as the page’s provenance.
Goals:
In this master’s thesis, the student will investigate a method to trace the provenance of Web pages
and the parts of which they composed. To achieve this, an extensive literature study will be
performed to identify existing approaches that might serve as a baseline for this purpose. The
student will then devise an improvement on these approaches, thereby exposing the full provenance
of a web page. This provenance can then be interpreted by a reasoner to suggest a trust
recommendation to the end user: a human being looking at the web page through a browser. By
working with standards from the W3C Open Web Platform, the solution devised in this thesis
potentially has a worldwide impact.
Bringing time-travel to data on the Web with efficient indexes
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Miel Vander Sande, Laurens De Vocht
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1 or 2
Number of theses: 1
Contact: Miel Vander Sande
Keywords: Data structures, Web APIs, Indexing, RDF, Semantic Web, Linked Data
Location: home, Zuiderpoort
Problem definition:
The Web of Data is an interconnected network of Linked Datasets residing on the Web. In
contrast to documents, concepts within the data are linked in a single global data space. These data
are modelled as a graph using the RDF framework, which describes data as triples (subjectpredicate-object). Although reading infrastructure has evolved significantly, writing this dataspace
is still an unsolved problem. How do you maintain such a dataspace where different users create,
read, update and delete statements? How do you take into account different views on the same
data? These problems pose new challenges for data storage, more specifically indexes, where
changes need to be remembered. Databases uses many different indexing algorithms like B-trees,
B+ trees, skip lists, hash tables and storage structures like Log-structured storage to enable fast and
concurrent read/write access. How do they perform for RDF? Can they be exploited by version
control systems?
Goals:
In this thesis, you initiate a building block for the Read/Write Web of Data. You dive into the
literature about general and versioned indexing strategies in RDF databases. Based on this
knowledge, you propose a technique that supports (a) fast triple pattern based retrieval, (b)
acceptable insertion/removal of triples and (c) track changes in order to retrieve past views. Finally,
the approach is implemented and evaluated using a use case, in order to verify the features
mentioned above.
Monitoring Science Related Conversations on the Web in Real-Time
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Laurens De Vocht, Anastasia Dimou
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Laurens De Vocht
Keywords: Web 2.0, Science 2.0, Researchers, Web of Data, Collaboration Tools, Social Media
Location: home, Zuiderpoort
Problem definition:
Digital libraries and online journals (such as IEEE, ACM) all have search engines to help scholars
find interesting resources. However, these approaches are often ineffective, mostly because
scholars: (i) only look-up resources based, at best, on their topics or keywords, not taking into
account the specific context and the scholar's profile; (ii) are restricted to resources from a single
origin. Of course, aggregators exist that index resources from multiple sources. The challenge is
therefore in matching research needs and contexts to opportunities from multiple, heterogeneous
sources. In other words, we should make the most of the wealth of resources for research through
relating and matching their scholar profile with the online available resources, publications and other
scholar's profiles.
Goals:
Combine streams of Web Collaboration Tools (e.g. Researchgate, Mendeley…) and Social Media
(e.g. Twitter, LinkedIn...) to track scientifically related communication and align it with the Web of
Data (such as COLINDA, DBLP, PubMed). This allows developing an efficient real-time monitor and
a useful environment for researchers. The monitor needs to allow interaction with the users, like a
dashboard. The user's personal research library and preferences could be matched with those of
others. This allows links to be made to social and research data beyond a single researcher's scope
and be a great source for inspiration (what is relevant to me?) and overview (what's hot right now
around me?). This should lead to more fine-grained details facilitating researchers to obtain a
sophisticated selection and linking of contributed resources based on previous assessments and
explored links.
Towards Intelligent Web Experiences: A Contextual User Journey
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Laurens De Vocht
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Laurens De Vocht
Keywords: Storytelling, Information Retrieval, Big Data, Linked Data, Pathfinding, User Interaction
Location: home, Zuiderpoort
Problem definition:
The number of resources that are now available to users on the Web, is rapidly expanding. Although
this Big Data may be structured and even linked, the underlying structure still looks like a maze to
most users. Therefore, there is an abundance of apps and web pages trying to hide this for the user
and of course there are search engines helping the user travel to the right page at all times.
Rather than imposing a linear journey on users through the hierarchies that website navigation and
lists of links impose, a networking-based contextual experience is based on a user’s search and the
relationships that form around what a user searches for. In this way, users navigate the network in
an order of their choice—establishing their own personalized web experience. There is still some
order as resources in the network are related to each other through modeled relationships and
ontologies. But, these relationships can run in multiple directions rather than one direction. The
categories become less important and the focus is put around the user and the content that is most
relevant to them.
You will investigate an approach to influence the generation of explanations of how resources are
related to each other in real-time in a personalized user context. Each explanation can be seen as a
useful/relevant re-combination of multiple associations between resources, for example:
 Trivia Finding / Personalized Story: DBpedia (Structured version of Wikipedia)
 Research 2.0: Recreate Events (Conferences) based on data of Web Collaboration Tools,
Digital Libraries, Linked Open Data
 Medicine: Drug discovery and genome data analysis.
Goals:
In this master’s thesis, the goal is to generate an experience which is both relevant to the user and
coherent. This includes an optimization on how the users interact with data, without violating any
rules of the context in which it is applied (e.g. chosen topics, required resources). Rather than living
in the past, you will investigate methods that make it possible to look toward the future, providing
inspiration as you discover things you didn’t know before. Relationships in the data can suggest new
songs you may want to listen to or people you may want to meet. The user's journey on the Web
evolves from being enforced linearly (passing through search engines over and over) to a network
of data represented in way they like.
User experience modeling in mobile applications using behavior stream visualization,
clustering and sentiment extraction
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Dieter De Witte, Azarakhsh Jalalvand, Paul Davies (UXProbe)
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: dieter.dewitte@ugent.be, Azarakhsh.Jalalvand@ugent.be
Keywords: Big Data Visualizations, Clustering, Sentiment extraction, Machine Learning
Location: Home, Zuiderpoort, Boerentoren Antwerp
Problem definition:
Nowadays, being a successful mobile application developer is a challenging task. Apart from the
fierce competition, you need to convince users to install your product and to keep using it. In this
context, it is of critical importance to identify and resolve issues/nuisances as soon as possible.
Therefore developers are eager to track user behavior and user sentiment in their application and to
identify, analyze and even predict bad user experiences.
UXProbe (http://www.uxprobe.org/), a startup specialized in usability design, created a service
which developers can integrate in their mobile application. The service collects event streams of
user interactions, errors that occur etc. What makes their service unique is that these event streams
are enriched with user generated sentiment feedback in the form of ratings, emoticons and minisurveys resulting in a very rich dataset. Deriving and communicating insights from this data in a
scalable fashion for increasing user bases, within limited time constraints is a challenging task for
which UXProbe needs you!
Goals:
As a first step you will create a number of dynamic dashboard-like visualizations to allow the visual
exploration of the user behavior patterns and the emotions associated with them. This will enable
development teams to quickly spot unexpected interaction patterns and assess the effectiveness of
the solutions they provide to address these.
In a second phase the thesis will, depending on your interests focus on:
Clustering the events into categories of ‘error motifs’.
The design of a prioritization queue based on the results of user sentiment analysis.
The design of a learning algorithm which can predict upfront which problems a user is likely to face
in the near future.
Analyzing and exploring links in biomedical graphs for drug discovery
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Dieter De Witte, Laurens Devocht & Filip Pattyn (Ontoforce)
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: dieter.dewitte@ugent.be, laurens.devocht@ugent.be
Keywords: Biomedical Data Mining, Graph Algorithms, Big Data
Location: home, Zuiderpoort
Problem definition:
Ontoforce is a startup that has created a solution to transparently query many clinical data sources
simultaneously. Their semantic technology - disQover - can query data from molecular databases,
scientific publications, clinical trials etc., allowing pharmaceutical companies to shorten their drug
discovery pipeline significantly. While the focus is currently on semantic technologies, there are
many questions which are more easily mapped onto pathfinding algorithms than to queries, a simple
example being: is there a path between lung cancer and smoking via genetic mutations? Neo4J is a
graph database optimized for graph traversal queries. Rik Van Bruggen, regional director of Neo
Technology in Belgium will be available as technical advisor on how to implement certain use cases.
Goals:
Design a number of specialized path finding algorithms in Neo4J and assess the
limitations/complementarity of both semantic technologies and graph databases for the use cases
presented. Investigate whether the algorithms can be translated to different context, for example in
automatic storytelling (www.everythingisconnected.be), which under the hood also relies on path
finding algorithms.
Turning Social Media Data to a Semantic Gold Mine
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Anastasia Dimou, Laurens De Vocht
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Anastasia Dimou
Keywords: Social Media, Linked Data mapping, Linked Data publishing
Location: home, Zuiderpoort
Problem definition:
Incorporating data from social media, e.g. twitter, Facebook, LinkedIn, into the Linked Open Data
cloud has a lot of value for marketing analysis, business and research. In real-world situations, this
information cannot be efficiently searched as we cannot query it, e.g. we cannot search when a user
last mentioned a place.
Goals:
In this thesis, you research extract-map-load approaches where handling the mapping is driven by
the data in real time. You investigate how to perform an efficient execution plan that handles such
data streams and maps them to RDF. The focus is on mapping social media data to the RDF data
model so you can bridge information from different social media in a way that you can identify
interesting trends around people, topics or events while you filter out the noisy data.
Visualizing biological sequence motifs using high performance multidimensional scaling
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Dieter De Witte
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: dieter.dewitte@ugent.be,
Keywords: Bioinformatics, Big Data Visualizations, Clustering
Location: home, Zuiderpoort
Problem definition:
The DNA of an organism consists of approximately 1% coding sequences, i.e. genes, and 99%
noncoding sequences. Regulatory sequences are hidden in the vicinity of the genes and help
regulate transcription (the process of translating DNA into RNA). In a later step RNA is then
synthesized into proteins. The accurate discovery of regulatory motifs in sequence data is very
difficult since these motifs are generally short and allow multiple wildcards.
An exhaustive discovery algorithm has already been developed which generates a database of
motifs which are overrepresented in sequences of closely related organisms. The amount of data is
however still too large to derive clear insights yet.
Goals:
The motif database generated by the discovery algorithm generates conservation information about
millions of short motifs. When one motif is found to be significant many highly similar motifs will
generally also be significant which implies that they can be clustered and the cluster as a whole
might have a biological meaning.
Multidimensional scaling is a technique which allows the visualization of high dimensional data by
mapping it onto a 2D or 3D Euclidean space while only requiring a well-chosen string distance
between the motifs.
The thesis student will investigate whether this algorithm can be used for large datasets of short
motifs, investigate the scalability and develop an end-to-end solution in which a biologist can
explore the data in an interactive way. As an extension the results from different motif algorithms
can be visually compared.
Low latency querying of huge Linked Medical datasets using Big Data technologies
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Dieter De Witte, Laurens Devocht
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: dieter.dewitte@ugent.be
Keywords: Big Data Architectures, Linked Data, Biomedical Data Mining
Location: home, Zuiderpoort
Problem definition:
The semantic web is a continuously expanding collection of linked datasets. Transparently querying
all these datasets together is a technical challenge for which multiple solutions have been
developed in the past years: from smart LDF clients to federated querying. If data dumps are
available another approach is possible: collect all dumps in a distributed file system (HDFS) and
query them using SQL-like tools developed for Big Data systems. Semantic data is generally
queried using the SPARQL language for which there is no real solution in the Big Data ecosystem
yet. Depending on the availability and the volume of the data this approach might be the preferred
or even the only solution.
Goals:
In a first step a literature study will be performed to get a clear overview of all attempts that have
been made to translate this problem into the Big Data space. In the next step a solution will be
developed which provides a SPARQL interface to a Big Data system of choice. Most recently
Apache Spark has risen to be the most popular big data technology and it would be interesting to
see if its superior data processing performance translates to linked data querying. As a benchmark
a set of queries on multiple medical datasets generated by the disQover platform of Ontoforce will
be used to compare the Big Data approach to the already available solutions based on federated
querying.
A journey planner for the galaxy
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Pieter Colpaert, Ruben Verborgh
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: pieter.colpaert@ugent.be
Keywords: hypermedia, REST, Linked Data, route planning
Location: home, Zuiderpoort
Problem definition:
Today, it would be impossible to write an app which plans routes throughout the entire galaxy: all
the transit data of e.g., buses, space shuttles, trains, bikes or space elevators would have to be
collected on one machine. This machine would then have to calculate hard to cache route planning
advice for all the user agents in the galaxy.
Within our lab we are developing a different server-client mechanism: instead of opening route
planning APIs for wide use, we suggest publishing the raw arrivals and departures using a REST
API. This way, route planning user agents can follow links to get more data on the fly. We are
implementing this data publishing interface with all the transit data that is already available worldwide and we have written a proof of concept user agent which now calculates the routes by
following links, and this within a reasonable query time.
Goals:
The goal of this thesis is to research the possibilities to make these user agents (which may be both
servers or mobile apps) more intelligent in various ways: e.g., the process could be sped up if we
can pre-fetch data on the client-side or the result can be more suitable for end-users if our useragent is able to discover other datasets (e.g., when I’m in Caracas, I only want these routes with the
least criminality reported, or when I’m in a wheelchair, I only want wheelchair accessible routes).
The user agent that is written needs to be generic: anywhere in the galaxy, it should be able to
automatically discover the right datasets published on the Web.
Building a Big Data streaming architecture for real-time Twitter annotation and querying
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Dieter De Witte, Frederic Godin, Gerald Haesendonck
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics,
Number of students: 1
Number of theses: 1
Contact: dieter.dewitte@ugent.be
Keywords: stream querying, linked data, neural networks, semantic annotation
Location: home, Zuiderpoort, Technicum
Problem definition:
Twitter is an online social network service that allows users to send short messages, aka tweets.
Currently over 500 million tweets are being generated every day. Automatically extracting
information from these tweets is a challenging task since they are short, can contain multiple
languages, contain spelling errors etc. Extracting information about tweets which are semantically
related, i.e. deal with a certain topic is far from trivial if they do not contain the same terms.
Goals:
In this thesis the student will be involved in setting up a streaming architecture for enriching tweets
with semantic information. Neural networks will be trained and used to label the tweets and to spot
named entities. The enriched stream will then be converted into semantic RDF triples which can be
queried using a streaming variant of the SPARQL language, for example C-SPARQL. Spark is a Big
Data technology stack that contains tools for both stream processing and batch analysis and is the
recommended technology for tackling this kind of problem.
Distributed query answering on the open Web
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Ruben Verborgh, Miel Vander Sande
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Ruben Verborgh
Keywords: Linked Data, Web, Semantic Web, querying, distributed systems
Location: home, Zuiderpoort
Problem definition:
Mail ruben.verborgh@ugent.be to discuss this subject.
What do your friends think of movies directed by Martin Scorsese?
What Nvidia graphics cards have few bug reports on Linux?
Is it cheaper to buy certain flights and hotels separately or together?
None of the above questions can be answered by a single data source, yet today’s Web technology
still focuses on single-source answer systems. This is problematic because a) it’s not scalable,
since that single source will need to process a lot of queries, and b) that source doesn’t have all the
data it needs to answer questions such as the above. The idea behind Linked Data Fragments
(http://linkeddatafragments.org/) is that clients, instead of servers, should answer queries. Sources
should offer fragments of data in such a way that clients can combine them to answer questions that
span multiple datasets.
A
client
that
works
with
a
single
data
source
already
exists
(https://github.com/LinkedDataFragments/Client). Your task in this master’s thesis is to extend this
client – or build a new one – so that it can query different data sources for a single query.
Goals:
Developing a scalable method to answer queries using different data sources.
Describing to a client which data sources are relevant for a given query.
Evaluating your solution on aspects such as accuracy and performance.
Querying multimedia data on the (social) Web
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Ruben Verborgh, Miel Vander Sande
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Ruben Verborgh
Keywords: Linked Data, Web, Semantic Web, querying, multimedia, images, video
Location: home, Zuiderpoort
Problem definition:
Mail ruben.verborgh@ugent.be to discuss this subject.
How would you find YouTube movies about “New York” in which people mention the Twin
Towers?
How could you find images that depict two people shaking hands?
Even though there is a large amount of metadata available on the Web, finding images and video
can be quite difficult. The goal of this thesis is to build an intelligent client (for instance, as a browser
extension) that is able to find multimedia items on the Web. This saves users many search
operations on different datasets. For this, you will need to combine metadata from different sources.
A starting point for this is the Linked Data Fragments client (http://linkeddatafragments.org/), which
already allows to query the Web Of Data. Your task is to blur the border between textual and
multimedia search, making it easier to find those media items users are looking for.
Goals:
Developing a client to find multimedia data on the Web.
Finding methods to query existing multimedia platforms such as YouTube and Instagram.
Evaluating your solution on aspects such as recall, precision, and performance.
Real-time querying of transport data on the Web
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Ruben Verborgh, Pieter Colpaert
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Ruben Verborgh
Keywords: Linked Data, Web, Semantic Web, querying, transport, train
Location: home, Zuiderpoort
Problem definition:
Mail ruben.verborgh@ugent.be to discuss this subject.
The Belgian rail website allows you to plan your journey, but only in very rigid ways. It does not take
into account your current location and plans. Suppose you need to be in a certain building in
Brussels for a meeting. That morning, you decide to take the train at 14:06. Unfortunately, that train
is severely delayed later on, but you won’t know that until you check the website again. In this thesis,
you develop a querying system over the Web that allows to retrieve real-time results that are
continuously updated. Based on data from different sources, your system automatically picks those
fragments that are necessary for users to plan their journey. You can build on existing work for Web
querying, such as Linked Data Fragments (http://linkeddatafragments.org/).
Goals:
Developing a real-time querying mechanism for transport data.
Planning a route using different fragments of data from the Web.
Evaluating your solution on aspects such as bandwidth and performance.
Predictive analytics for the Internet of Things
Promotors: Erik Mannens and Rik Van de Walle
Supervisors: Dieter De Witte, Sven Beauprez (IoTBE.org)
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 2
Number of theses: 2
Contact: dieter.dewitte@ugent.be
Keywords: Internet of things, linked data publishing, stream reasoning, big data analytics
Location: home, Zuiderpoort
Problem definition:
In the upcoming years the number of devices for the Internet of Things will grow exponentially. IoT
will therefore become the largest source of streaming data. In order to derive insights from this data
it should be converted into a format which can be queried and allows easy semantic enrichment.
Linked open data is the ideal candidate to fulfill this task. To prepare for this innovation a prototype
environment is required which will reveal the challenges for the upcoming data explosion. As a data
source we will make use of Tessel Microcontrollers (www.tessel.io) which can be easily configured
using only JavaScript and which are extendible with a wide range of sensors: audio, video,
temperature, Bluetooth etc. The built-in Wi-Fi allows for a straightforward data transfer to a Big Data
infrastructure.
Goals:
In this thesis the student(s) will be involved in setting up an architecture for analyzing and enriching
sensor streams with semantic information. The annotated streams can be queried using a streaming
variant of the SPARQL language, used for (static) linked data.
In a first phase the student will build an experimental setup with Tessel devices as the data source.
The data generated by the devices will be automatically ingested in a Big Data streaming
architecture. For the analytics platform the student will explore the possibilities of the Apache Spark
stack which contains tools for both stream processing and batch analysis.
Since multiple students can work on this topic the focus of the thesis can be aligned with the
interests of the student: if the data streams are captured he/she can focus on enrichment,
visualizations or on optimizing the performance of the data pipeline.
Web-based framework for real-time data visualization and scene graph management
Promotor: Peter Lambert and Rik Van de Walle
Supervisors: Jelle Van Campen, Tom Pardidaens and Christophe Herreman (Barco)
Study Programme: Master in Computer Science Engineering, Master of Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Christophe Herreman (Barco)
Keywords: html5, web, real-time, scene graph, data visualization
Location: home, Zuiderpoort, Barco
Problem definition:
Barco develops systems and solutions for professional visualization applications in many different
markets and increasingly requires web-based front-ends for these applications. The applications
typically have stringent requirements towards image quality, latency and user experience and deal
with large amounts of live video and data channels. At this moment, front-end applications are
developed by integrators or specifically for a single customer.
Goals:
The goal of this project is to develop prototypes, which demonstrate different technologies to render
highly dynamic scenes such as geographical maps or 3D city models, and augment the visualization
with real-time video and meta-data. After analyzing the requirements from various divisions, a study
of the state-of-the-art should be performed including evaluating available technology (commercial or
open source libraries). The suitability of the libraries is to be explored with regards to performance,
latency, scalability and feature completeness. Then, a framework and software architecture is to be
designed and implemented to enhance these capabilities with the in-house capabilities for real-time
video and meta-data visualization and also to enable handling different types of content in single
browser window, possibly combining the capabilities of several libraries.
This thesis is jointly supervised with Barco (www.barco.be).
Designing an Ontology-Driven Dialog System for Natural Language Understanding
Promotoren: Kris Demuynck en Erik Mannens
Supervisors: Gaëtan Martens(Gaetan.Martens@nuance.com) en Azarakhsh Jalalvand
Study Programme: Master Computer Science Engineering, Master Electrical Engineering, Master
Mathematical Informatics
Number of students: 2
Number of theses: 1
Contact: Gaëtan Martens (gaetan.martens@nuance.com), Azarakhsh Jalalvand
Keywords: Automatic Speech Recognition, Natural Language Understanding, Semantic Web,
Dialog System
Location: home, Zuiderpoort
Dialog Management (DM) and Natural Language Understanding (NLU) are fields of computer
science, artificial intelligence, and linguistics concerned with the interactions between computers
and human (natural) languages. Nuance Communications is the market leader in those
technologies and delivers currently the most significant advancements in speech recognition
technology.
Problem definition:
An Automatic Speech Recognition (ASR) + NLU system transforms an input speech signal into a
semantically enriched output which consists of intents and interpretations. A dialog system (DS) is
an engine responsible for DM, which allows having a conversation with a machine based on a
predefined set of concepts. These concepts and their relationships are modeled using Semantic
Web technologies by means of an ontology. This ontology defines the behavior of the DS relying on
a semantic reasoner. For example, if the user says “I want to make a phone call” (intent=“Phone”)
then the DS should ask for additional information such as: “Do you want to call somebody from your
contact list or do you want to dial a number?” On the other hand if the user said “I want to call
Gaëtan Martens on his cell phone” the system should not ask for additional information since the
required interpretations (i.e., contactName=“Gaëtan Martens” and phoneType=“cell phone”) are
already available.
Goals:
In this Master thesis the students will build a DS which relies on an OWL ontology to define its
behavior. This ontology has to be created based on a set of uses cases provided by Nuance. The
ontology has to model the intents, the concepts (i.e., the corresponding NLU interpretations), and
the relationships between the concepts. The next step is then to build the DS around an existing
semantic reasoner. The final goal of this challenging thesis is to have a functional DS, built using
open source libraries, that is able to use the output from Nuance’s state-of-the-art ASR+NLU system
and is configured with the created ontology.
The students will have the opportunity to work on cutting-edge technology within the Nuance’s
automotive R&D team in an international context.
Building a Natural Language Dialog System with Semantic Web technology
Promotoren: Kris Demuynck en Erik Mannens
Supervisors: Gaëtan Martens(Gaetan.Martens@nuance.com) en Azarakhsh Jalalvand
Study Programme: Master Computer Science Engineering, Master Electrical Engineering, Master
Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Gaëtan Martens (gaetan.martens@nuance.com), Azarakhsh Jalalvand
Keywords: Automatic Speech Recognition, Natural Language Understanding, Semantic Web,
Dialog System
Dialog Management (DM) and Natural Language Understanding (NLU) are fields of computer
science, artificial intelligence, and linguistics concerned with the interactions between computers
and human (natural) languages. Nuance Communications is the market leader in those
technologies and delivers currently the most significant advancements in speech recognition
technology.
Problem definition:
An Automatic Speech Recognition (ASR) + NLU system transforms the input speech signal into a
semantically enriched textual output. This output consists of an n-best list of intents (e.g., “Call”) with
a set of possible interpretations (such as contactName=“Gaëtan Martens” and phoneType=“cell
phone”) and corresponding probabilities. The dialog system (DS) is the engine responsible for DM,
which allows having a conversation with a machine based on a predefined set of concepts. This set
of concepts (and their inter-relationships) is modeled using Semantic Web technologies by means of
an ontology. This ontology then defines the behavior of the DS. The DS is able, given its current
state, to recognize an intent or ask for more information when the input is ambiguous by applying
reasoning techniques.
Goals:
In this master thesis, the goal is to link an existing ASR + NLU system with a DS via ontologies. At
first, the supported intents and interpretations of the NLU system must be formally defined in the
ontology and missing intents (required by the DS) will have to be added. Further alignment tasks will
have to be investigated by testing and iterative enhancements such as improving the mechanism of
the DS to take into account the NLU probabilities. The final goal is to have a functional speechenabled ontology-based end-to-end DS, which is still a challenge given the current state of the art.
The student will have the opportunity to work on cutting-edge technology within the Nuance’s
automotive R&D team in an international context.
The Multi-sensory and Sentiment-aware eBook
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Hajar Ghaem Sigarchian, Tom De Nies, Frank Salliau
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact:Hajar Ghaem Sigarchain
Keywords: e-Text-Book, Sentiment Analysis, Internet Of Things, Semantic Web
Location: home, Zuiderpoort
Problem definition:
People expect more and more when reading e-Books, in terms of entertainment and immersion.
The recent advances in smart environments theoretically allow people to experience an eBook not
only using their eyes but also using their other senses, such as hearing, touch (e.g., thermoception),
etc. Content of such enhanced e-Books can become augmented and reacted upon by smart objects,
using the Internet of Things (IoT). However, many challenges remain to make this a reality, such as
the automatic annotation, exchange of data, and timing (e.g., should the room react any time a
person flips a page of the eBook?).
Goals:
The goal of this master’s thesis is create digital books as an object in IoT. The student needs to
perform a literature review, in order to have better understanding of the concepts and current state
of the art. In addition he/she will propose an appropriate architecture, and data representation
format. The research domains will include smart rooms, sensors, Internet of Things, digital
publishing and Semantic Web. The next step is implementing a prototype as a proof-of-concept.
Eventually, he/she needs to test the prototype to evaluate and qualify the relevancy of the solution
to the end-user (the reader of the book).
Using Events to Connect Books Based on Their Actual Content
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Pieter Heyvaert, Ben De Meester, Frank Salliau
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Pieter Heyvaert
Keywords: books, events, Semantic Web, media, metadata
Location: home, Zuiderpoort
Problem definition:
Many books tell a story about a certain event or a series of events. Different stories (in multiple
books) might even contain the same events. An example of such a series of events can be the
Battle of Waterloo, which includes separate events for the preparation, the first French attack, the
capture of La Haye Sainte, the attack of the Imperial Guard, the capture of Plancenoit, and so on.
Currently, the metadata extracted from books mostly denotes topics, known locations, people, dates,
and so on. However, the connection between people, their location and the date is not considered,
and these elements are what an event consists of. When having this information available, one is
able to determine the similarity between different books regarding a specific event. This is not only
limited to books: when event information is also available for movies, TV series, news articles, and
so on, one is also able to determine how the literature/movies interpret real events: what happens
when non-fiction (e.g., newspaper articles) takes a fiction approach (via e.g., movies) is one of the
questions that might be answered here.
Goals:
You will need (1) to determine how events can be dissected from the several types of media, with
the focus on books, (2) to decide on the granularity of the events extracted from the media, (3) to
determine how to incorporate existing information, if any, about the events and media, and (4) to
determine/visualize the events found in the different sources. As part of the thesis you will need to
create a prototype which takes media (such as books) as input and outputs the events. Next, you’ll
need to visualize the connection between the different media based on the found events. Also the
difference between events can be determined: two books might talk about the same big event,
however, the events making up the big event may be different, may be in a different order, and so
on.
Deriving Semantics from Styling
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Ben De Meester
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Ben De Meester
Keywords: HTML5, Semantics, CSS
Location: home, Zuiderpoort
Problem definition:
HTML is the de facto publication format for the World Wide Web for humans. Using styling
properties (e.g., in a CSS stylesheet), it is possible to style the HTML document to make it pretty.
However, the HTML file itself is rarely pretty for machines. Coding such as ‘<p class=”title”>’ is a
terror, and contradictory to the semantic structure that HTML5 can provide. One thing that will
always be consistent however, is the visual looks of an HTML page versus its intended semantics.
E.g., ‘<p class=”title”>’ will look similar to ‘<h1>’ for a user, but the latter has a lot better semantic
meaning than the former. And better semantics means better processing of the HTML, which results
in a better Web.
Goals:
The goal of this thesis is to improve existing HTML pages, their structure and semantic meaning, by
using, among others, its visual characteristics. To that end, the student needs to define which
criteria could be important, e.g., not only font-size could be important, but also font-weight, position
on a page, name of the CSS class, etc. Next, the student should provide for a proof-of-concept that
can effectively recognize and improve bad HTML constructs. Many techniques may be used,
involving clustering, Natural Language Processing, and ad hoc processing instructions.
Trusting query results on the open Web
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Ruben Verborgh, Tom De Nies
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Ruben Verborgh
Keywords: Linked Data, provenance, trust, Semantic Web, querying, Web, HTTP, client/server
Location: home, Zuiderpoort
Problem definition:
Mail ruben.verborgh@ugent.be to discuss this subject.
The Web way of answering a question is to find combined answers from different sources. But
how can we be sure that the final answer is based on sources we can trust? This is the question
you will answer in this thesis.
Because information is spread across different places, the old database paradigm of “query =>
answer” doesn’t really work anymore on the Web. Linked Data Fragments
(http://linkeddatafragments.org/) capture this idea: a client asks servers for different parts of
information and is able to combine it by itself. We built a Node.js application that can query the Web
in such a way (https://github.com/LinkedDataFragments/Client). What this client doesn’t tell you
(yet), is where the different parts of the answer come from; it doesn’t give you a guarantee that the
answer is correct/trustworthy. By combining Linked Data Fragments with provenance technology,
we can precisely assess the trustworthiness of data.
Goals:
- Developing a method to combine the trust in different knowledge sources.
- Describing the trust in an answer that encompasses different sources.
- Developing a client that queries the Web and gives an answer you can trust.
Automatic Composition of Context-based Content in Digital Books
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Hajar Ghaem Sigarchian, Tom De Nies, Wesley De Neve, Frank Salliau
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Hajar Ghaem Sigarchain
Keywords: e-Text-Book, Widgets, Ubiquitous environment, Semantic Web, Profile Manager
Location: home, Zuiderpoort
Problem definition:
Books, even digital books are a static medium, with fixed contents. There is currently no means of
personalizing and tailoring books to the person who reads them. However, the technology to provide
dynamic contents inside a digital book already exists. Depending on contextual criteria such as the
reader’s interests, location, language, culture, the contents can adapt itself to fit the reader’s needs.
As a general use case, we can refer to non-fiction books such as tourist guide books that e.g.
automatically update their contents based on the reader’s geolocation.
Educational textbooks also provide an excellent use case to prove this principle, with the textbook
adapting itself to the student’s progress and changing interests. Publishers will also benefit from this
approach: it facilitates reuse of existing content; it also can lead to a significant reduction of both
creation and distribution costs and even lead to new business models.
Goals:
The goal of this master’s thesis is to automatically change digital books based on their context with
external open content. The student needs to do a literature review, in order to have better
understanding of the concepts and current state of the art. In addition he/she will propose an
appropriate architecture, and data representation format. The research domains will include
mashups, widgets, cloud-based synchronization, versioning and the Semantic Web.
The next step is implementing a prototype as a proof-of-concept. Eventually, he/she needs to test
the prototype to evaluate and qualify the relevancy of the automatically provided contents.
Mining anonymized user data for predictive analysis
Promotor: Sofie Van Hoecke and Peter Lambert
Supervisors: Glenn Van Wallendael, Benoit Marion (Benoit.Marion@tpvision.com), Dirk Van Gheel
(Dirk.VanGheel@tpvision.com)
Study Programme: Master Computer Science Engineering, Master Industrial Engineering:
Elektronica – ICT, Master Industrial Engineering: Informatica
Number of students: 1
Number of theses: 1
Contact: Sofie Van Hoecke
Keywords: Data mining, recommendations, crash prevention
Location: home, Zuiderpoort, TPVision Zwijnaarde
Problem definition:
To continuously improve the quality of Philips televisions, TPVision incorporates anonymized user
logging on Android enabled Smart TVs. However, as current TVs are complex systems, a large
amount of user data is logged. To improve the evaluation of user data, advanced data mining
techniques are required in many areas of our operations.
Efficient data mining can result in an optimized recommendation engine to recommend more
interesting TV channels or Android apps, improving the user’s lean back experience. Additionally,
watch later lists combining YouTube, Vimeo, and Spotify can be optimized by providing a prioritized
list.
Furthermore, from the user data the market impact of certain features (such as the availability of
broadcast tuners, pre-installed apps, DLNA connectivity etc.) can be predicted. Identifying
relationships between user behavior and interaction with those features taking into account general
user clusters and regional differences can result in differential marketing strategies.
Finally, the performance of the TV can be improved using machine learning techniques for analysis
of logged data and intervene in a timely manner to prevent a potentially hazardous crash to happen.
Goals:
The goal of this master’s thesis is to apply data mining techniques on anonymous user data from
currently deployed Philips Android TVs within Europe.
Topics to improve user experience, technical enhancements or commercially interesting research
are considerable. Depending on the student’s interest, the actual problem statement can be finetuned.
TP Vision is a dedicated TV player in the world of visual digital entertainment. TP Vision is part of
TPV Technology, the #1 monitor manufacturer in the world. At its Innovation Site Europe, Ghent we
design televisions of the future, for Philips and other brands. NetTV, Ambilight, Android TV and
Cinema 21:9 have all been developed in this innovative and driven environment. Recognition of our
activities is visible in numerous awards such as prestigious EISA-awards.
This master’s thesis is within the Advanced Software Development department which drives the
innovation chain from conceptualization to feasibility and prototyping, as well as leads technology
and standardization roadmaps.
A Proof checker for the Semantic Web
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Dörthe Arndt, Ruben Verborgh
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Dörthe Arndt
Keywords: N3Logic, proof, reasoning, semantic web
Location: home, Zuiderpoort
Problem definition:
The Semantic Web enables computers to understand and reason about public data. There exists a
huge number of reasoners which are able to draw conclusions based on common knowledge (e.g.
Pellet, CWM or EYE). Some of them also provide proofs to clarify how they came to a certain result.
But what is a proof? How can we be sure that a proof is correct? When do we trust it? Could they
be lying?
An independent checker is needed!
Goals:
With this thesis you will help to find a solution for this problem: You will get a better understanding of
N3-logic, the logic used by reasoners such as EYE or CWM
((http://www.w3.org/2000/10/swap/doc/cwm.html, http://eulersharp.sourceforge.net/).
You’ll learn what the current proof format of these reasoners looks like and how it could be improved.
This knowledge will enable you to implement an independent proof checker for N3-proofs which can
handle proofs written in the N3-proof ontology (http://www.w3.org/2000/10/swap/reason.n3).
On-site tijdsregistratie en efficiëntie-analyse in het operatiekwartier
Promotor: Sofie Van Hoecke, Patrick Wouters
Supervisors: dr. Piet Wyffels (UZ Gent)
Study Programme: Master of Science in de industriële wetenschappen: elektronica-ICT - Campus
Kortrijk, Master of Science in Computer Science Engineering, Master of Science in de industriële
wetenschappen: elektronica-ICT - Campus Schoonmeersen, Master of Science in de industriële
wetenschappen: informatica
Number of students: 1
Number of theses: 1
Contact: Sofie Van Hoecke
Keywords:
Location: home, Zuiderpoort, Technicum or UZ Gent (K12)
Problem definition:
Het organiseren en uitvoeren van een operatieprogramma binnen de dagelijks toegemeten werktijd
blijkt één van de meest complexe taken binnen een ziekenhuis. Er bestaat heel wat literatuur waarin
werkmodellen en systemen worden voorgesteld maar deze zijn slechts beperkt toepasselijk, o.a.
door belangrijke internationale verschillen in de structuur van de gezondheidszorg. Een meer recent
en universeel probleem dat de werkbaarheid van gangbare organisatiemodellen ondermijnt is het
spanningsveld tussen opgelegde budgettaire beperkingen en de stijgende zorgvraag.
Het enige afdoende antwoord hierop is een efficiëntere inzet van bestaande middelen en personeel
en een optimalisatie van de workflow (planning en uitvoer) in deze multidisciplinaire en
hoogtechnologische omgeving. Dit is geen eenvoudige opdracht gezien de aanwezigheid van
onvoorspelbare factoren (urgenties, complicaties), de betrokkenheid en uiteenlopende belangen
van verschillende disciplines (chirurgen, anesthesisten, verpleegkundigen, technici) en de
afstemming op diensten die patiënten moeten aanleveren en ontvangen.
Goals:
Het doel van deze masterproef is het ontwerpen van een intelligent, flexibel en gebruiksvriendelijk
systeem (bv. onder de vorm van web-gebaseerde toepassing) dat een accurate tijdsregistratie op
de werkvloer mogelijk maakt met aandacht voor de verschillende fases en deelaspecten die aan
bod komen bij, en invloed hebben op het verloop van een chirurgische ingreep (opstart tijd, turnover
tijd, installatietijd...). Het ontwikkelen van een geschikt datamodel is hierbij heel belangrijk. De data
die aldus opgemeten wordt, moet eenvoudig en efficiënt kunnen bevraagd worden om
tijdsverliesposten in kaart te brengen en na root cause analysis systematische verbeterinitiatieven te
lanceren.
Tijdens deze masterproef zal de student kennis opdoen omtrent de werking van een
operatiekwartier en met de verschillende partijen afstemmen om de verschillende standpunten en
invalshoeken op te nemen in het ontwerp. Er bestaat momenteel een grote en zeer brede interesse
in (en markt voor) dergelijke toepassingen.
Deze masterproef loopt in samenwerking met de dienst Anesthesie van het UZ Gent.
Real-time data-enrichment and data-mining of multilingual databases in procurement
environment
Promotors: Erik Mannens & Rik Van de Walle
Supervisors: Ruben Verborgh & Dieter De Witte
Study Programme: Master Computer Science Engineering, Master Mathematical Informatics
Number of students: 1
Number of theses: 1
Contact: Manu Lindekens (GIS International)
Keywords: datamining, data-enrichment, data-cleaning
Location: GIS International – Stapelplein 70 – 9000 Gent
Problem definition:
In a B2B world we as a procurement service provider (www.gisinternational.net) do receive lots of
information from our customers on items we need to buy from different sort of vendors. This
information is in most cases very limited, in different languages, incomplete and polluted.
Based on that limited information we, in a first phase, categorize these articles into different
commodities. This categorization will allow us to define better which vendors to target. A second
step is to enrich the data to be more accurate in ordering the material. These are all steps that are
today done manually but are very labor-intensive. An off-the-shelf software package that does all
this is currently not available.
Secondly the data in our own ERP system needs to be cleaned and enriched in real-time. Currently
we have identified different parties to tackle this problem with intelligent software.
Goals:
The goal of this thesis is to do be the link between the different parties who will develop and
implement the software to solve this problem. Therefore a good understanding of the needs, the
processes and how the software will work is important. GIS International has the knowledge on the
parts, knows which information is needed and commercially where to get it from. The other parties
are top research and technology companies which are experts in their fields. Purpose is to build a
state-of-the-art add-on to the ERP software of GIS (MS Dynamics NAV) which will be the
differentiator in the market.
Do not hesitate to contact us when you are interested in this topic and have some questions!
An extensible web-based intermodal route planner for Belgium
Promotor: Erik Mannens and Rik Van de Walle
Supervisors: Pieter Colpaert, Ruben Verborgh
Study Programme: Industriële wetenschappen
Number of students: 1
Number of theses: 1
Contact: pieter.colpaert@ugent.be
Keywords: hypermedia, REST, Linked Data, route planning
Location: home, Zuiderpoort
Problem definition:
Developers of route planning apps today have to be satisfied with web-services which do the hard
work for them: they offer a black-box algorithm with a finite set of functionalities (e.g.,
http://api.myapp/?from=...&to=...). When the developer would like to make the algorithm take into
account other modes of transport or take into account different edge weights (e.g., depending on
wheelchair accessibility, criminality statistics or probability of being too late), it needs to request
these features to the server-admin.
At MultiMedia Lab, we are researching an alternate server-client trade-off: instead of exposing route
planning algorithms over HTTP, we suggest publishing the raw arrivals and departures using a
REST API with paged fragments (http://linkeddatafragments.org). This way, route planning user
agents, which now execute the algorithm on their own, can follow hypermedia controls to get more
data on the fly.
When we publish an ordered list of arrival/departures (connections), route planning can be done by
relaxing each connection once (shortest path in a DAG). We publish this graph as Linked Open
Data resources for route planning purposes.
Goals:
The goal of this master thesis is to research different strategies to merge graphs from different
transport
modes
on
the
scale
of
Belgium.
A
visualization
(cfr.
http://kevanahlquist.com/osm_pathfinding/) should be created to understand the different strategies
for merging different transport modes.