• No results found

Enhancing the Recommendations of NPO Start with Metadata

N/A
N/A
Protected

Academic year: 2021

Share "Enhancing the Recommendations of NPO Start with Metadata"

Copied!
14
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Enhancing the Recommendations of NPO Start

with Metadata

submitted in partial fulfillment for the degree of

master of science

Eileen Kapel

12369462

master information studies

data science

faculty of science

university of amsterdam

2019-06-28

External Supervisor External Supervisor

Title, Name Dr Maarten Marx Robbert van Waardhuizen

Affiliation UvA, FNWI, IvI NPO

(2)

ABSTRACT

The current recommendation system of the NPO Start service uses a collaborative filtering approach for serving out video recommenda-tions. However, a lot of available metadata is unused in this system that can be utilised for providing recommendations. This thesis aims to determine whether a hybrid recommendation system that utilises this metadata can perform better than the NPO Start recommenda-tion system. Based on experiments where interacrecommenda-tion informarecommenda-tion and different combinations of metadata features were supplied to a hybrid LightFM model, the model that utilised the metadata fea-tures broadcaster, description, genres and title performed similar but slightly higher than the model utilising no metadata features. After the hyperparameters of the former model were optimised and compared to the current model of the NPO Start recommendation system, it was concluded that a hybrid recommendation system us-ing metadata can perform better than the current recommendation system of NPO Start.

KEYWORDS

Recommendation systems; Hybrid recommendation systems; Meta-data

1

INTRODUCTION

1

Humans are tasked with making thousands of decisions daily that

2

may range from selecting what outfit to wear to which television

3

series to watch. Recommendations can make the process of these

4

decisions more efficient by sorting through potentially relevant

5

information and making recommendations customised to the

indi-6

vidual user [10, 13]. One system that employs recommendations is

7

NPO Start1. It is a video-on-demand service from the NPO (Dutch

8

Public Service Media) that gives people the ability to watch series,

9

movies and documentaries as well as watch live television online.

10

Personalised recommendations are available for registered users

11

of the service, which is based on a collaborative filtering approach.

12

However, there is a lot of metadata available about the offered

13

content that is unused in this current method. In this thesis, the

14

metadata of broadcasts will be utilised to determine if it can improve

15

the performance of the current video recommendation system.

16

This is achieved by answering the main research question: ’Can

17

a hybrid recommendation system using metadata perform better than

18

the current recommendation system of NPO Start which uses

collabo-19

rative filtering?’. The research is split into three sub-questions:

20

RQ.1 What is the performance of the NPO Start recommendation

21

system?

22

RQ.2 Which content features improve the performance of the

23

hybrid recommendation system the most?

24

RQ.3 Can the performance of the NPO Start recommendation

sys-25

tem be improved by implementing a hybrid recommendation

26

system?

27

The thesis is structured as follows: first, some background

infor-28

mation on the NPO Start service is provided in section 2. Various

29

1www.npostart.nl

related works of literature are outlined in section 3 that relate to the

30

goal of the thesis. The methodology employed during the research

31

is discussed in section 4 and is followed by the results in section

32

5. Subsequently, the conclusions of these results are presented in

33

section 6 after which a discussion of the choices and possible future

34

work is presented in section 7.

35

2

BACKGROUND

36

NPO Start is a service that offers users the ability to watch video

37

content on demand. This video content is displayed to users in

so-38

called “ribbons“ or rows that have a certain theme, like ‘Populair‘,

39

‘Nieuw‘ and ‘Aanbevolen voor jou‘. Each ribbon consists of a ranked

40

list of several items and an item represents a series that can be

41

streamed. A ribbon typically displays about five items and allows

42

for exploring more items that fit with that particular theme.

43

2.1

The NPO Start Recommendation System

44

Users of the service that have an account have the ability to receive

45

several personalised ribbons that contain items that are

recom-46

mended to a specific user. These recommendations are materialised

47

on the front page of the service in the two ribbons ‘Aanbevolen

48

voor jou‘ en ‘Probeer ook eens‘. This thesis focuses on the

‘Aan-49

bevolen voor jou‘ ribbon. The recommendations for this ribbon are

50

produced by a collaborative filtering approach that utilises the

his-51

tory of users their interaction information with items. These user

52

interactions are grouped on series level and evaluated by pairs of

53

series that are frequently watched together, or coincide often, with

54

the history of the user. Of these coincidences, the top 100 pairs are

55

extracted after they are ordered based on their frequency. A sliding

56

window technique of the past 21 days is used for the interaction

57

information, which is based on the intuition that recent events are

58

more relevant than older ones [1]. These recommendation lists are

59

updated hourly at peak times and bi-hourly off-peak.

60

3

RELATED LITERATURE

61

Recommendations are based on ratings that are explicitly given

62

by users or ratings that are implicitly inferred from users their

63

actions [13], like a click or a certain watch duration. There are

64

three main approaches for building a system that gives out

recom-65

mendations. The first approach utilises information about items

66

for a content-based system and recommends items that are

simi-67

lar to the well-rated items of a user. The second approach utilises

68

users their interaction information with items for a collaborative

69

filtering system and recommends items that are well-rated by other

70

users who have the same pattern of ratings. Lastly, there is a hybrid

71

recommendation system, which is a combination of the two

pre-72

vious approaches, that exploits item information and interaction

73

information to provide recommendations. The first two approaches

74

each have their own shortcomings, like overspecialisation, rating

75

sparsity and cold-start [1, 8], that hybrid systems aim to overcome

76

to provide more accurate recommendations.

77

In this section, an overview of the current state of hybrid

rec-78

ommendation research is given. Furthermore, a few personalised

(3)

services that employ recommendation systems are described and,

80

lastly, the representation of features in recommendation systems is

81

touched upon.

82

3.1

Hybrid Recommendation Systems

83

Hybrid recommendation systems are able to deal with scenarios

84

where there is little data on new users and items, also called the

85

cold-start problem, or when there is sparse interaction data

avail-86

able. This is achieved by joining content and collaborative data in

87

the system to produce recommendations that not only takes into

88

account similar users but also personal interests.

89

Several techniques exist for combining content-based and

collab-90

orative filtering systems of which the weighted, mixed, switching

91

and feature combination techniques [3] are most frequently used.

92

A weighted hybrid recommendation system is one where the rating

93

of an item is a combination of the content-based and collaborative

94

rating. Alternatively, a mixed hybrid recommendation system

out-95

puts items from the different approaches together. The switching

96

recommendation system uses a different approach dependent on

97

the situation, for example a content-based system could be used

98

when there is little interaction information and in other cases a

col-99

laborative filtering system is used. Finally, the feature combination

100

technique combines both content-based and collaborative

infor-101

mation into a single recommendation algorithm. This technique

102

causes the recommendation system to rely less on the amount of

103

ratings per item and allows for less-known but similar items to be

104

recommended.

105

Furthermore, different algorithms can be employed in the hybrid

106

recommendation techniques, however most techniques employ

ma-107

trix factorisation for the collaborative filtering part of the system.

108

This algorithm was popularised by the solution of the Netflix Prize

109

competition that employed matrix factorisation using the

alternat-110

ing least square (ALS) algorithm [2, 6, 7]. Some works that have

111

successfully employed matrix factorisation in their hybrid

recom-112

mendation systems are Rubtsov et al. [14], Ludewig et al. [11] and

113

Al-Ghossein et al. [1]. Rubtsov et al. used the feature combination

114

technique by making use of the LightFM library, which is a Python

115

implementation of a matrix factorisation model that can deal with

116

user and item information [8], paired with a weighted

approximate-117

rank pairwise loss. Ludewig et al. made use of matrix factorisation

118

in their model by combining it with a k-nearest-neighbour

tech-119

nique and using the ALS algorithm. Content was incorporated into

120

the model by weighing the matrix factorisation results with the

121

IDF (inverse document frequency) score of titles to produce the

122

final list of recommendations. Lastly, the feature combination

hy-123

brid recommendation system by Al-Ghossein et al. merged matrix

124

factorisation with topics extracted using topic modeling for online

125

recommendation.

126

Furthermore, neural networks have also been combined into

hy-127

brid recommendation systems due to its recent popularity. Volkovs

128

et al. [16] produced a two stage model for automatic playlist

contin-129

uation that first employs weighted regularised matrix factorisation

130

to retrieve a subset of candidates and then uses convolutional

neu-131

ral networks and neighbour-based models for detecting similar

132

patterns. In the second stage features of playlists and songs are

133

combined with the items after which the final ranking of

recom-134

mendations are produced. Another novel approach for automatic

135

playlist continuation is the weighted hybrid recommendation

sys-136

tem made up of a content-aware autoencoder and a character-level

137

convolutional neural network (charCNN) by Yang et al. [18]. The

138

content-aware autoencoder alternates in predicting artists fitting

139

in a playlist and playlists fitting with an artist. The charCNN takes

140

a sequence of characters as input, in this case a playlist title, and

141

predicts the most fitting tracks with this sequence. The output

142

of both components is linearly combined and produces the final

143

recommendations.

144

3.2

Personalised services

145

Recommendation systems are frequently employed by services to

146

provide a personalised experience to their users. This occurs in

147

different domains, e.g. product recommendation by Amazon, music

148

recommendation by Spotify and video recommendation by YouTube

149

and Netflix.

150

Netflix is a service where the whole experience is defined by

per-151

sonalisation [4]. This is primarily showcased on its homepage which

152

consists of rows of recommended videos with a similar theme, like

153

‘Top Picks‘ and ‘Trending Now‘, that are ranked by a personalised

154

video ranker. Two of these rows, namely the genre and because

155

you watched rows, take the content of the videos in account for the

156

recommendations. The videos in the genre row are produced by a

157

single algorithm that takes a subset of all videos that corresponds

158

to a specific genre. Examples of such rows are ‘Suspenseful Movies‘

159

and ‘Romantic TV Movies‘. The because you watched row bases

160

its recommendations on a single video that is watched by a user

161

and uses a video-video similarity algorithm. This algorithm is not

162

personalised, but the choice of which because you watched rows

163

are offered to a user is personalised. An example of this kind of row

164

is ‘Because you watched Black Mirror‘.

165

The streaming service Spotify employs personalisation in several

166

areas, like its homepage and the feature that allows for automatic

167

playlist continuation. The homepage allows users to discover new

168

playlists which are similar to the playlists and tracks a user has

169

previously interacted with. The automatic playlist continuation

170

feature adds one or more tracks to a playlist that fits the original

171

playlist of a user [19]. This feature takes not only the collaborative

172

information of playlists and their corresponding tracks into account

173

but also playlists their content in the form of titles and featured

174

artists.

175

3.3

Representation of Features

176

Multimedia content, like songs, films and series, is often represented

177

by a set of features. A feature is information that describes an

178

attribute of an item, like its title, plot, genre or release year.

179

Features of content are used in content-based recommendation

180

systems and thus also in hybrid recommendation systems. The

181

performance of such systems is predicated on the quality of the

182

features, meaning that features derived from high-quality metadata

183

lead to a better performance [8, 14]. This is evident in the work of

184

Soares & Viana [15] where the version of their recommendation

185

system that used more granular metadata as features, e.g. genres

186

and sub-genres, resulted in recommendations of a higher quality. If

187

(4)

high-quality metadata is not available then good quality metadata

188

can be obtained from item descriptions, like actor lists and synopses

189

[8]. However, metadata of a lower quality, e.g. by being sparse, may

190

result in overfitting and causes models to not make use of content

191

in an effective way [19].

192

Feature selection is frequently used to improve the quality of

193

metadata. It is a method where rather than using all the features

194

only a subset of the features is used [12]. By carefully selecting

195

this subset of features a better effectiveness of the system can be

196

achieved. For example, the hybrid recommendation system by Soar

197

et al. that only employed the director as feature, instead of all

198

the features, resulted in recommendations that were more precise

199

[15]. This is likely explained by the fact that a director can provide

200

specific information on the potential quality of content that cannot

201

be described with another set of metadata elements, e.g. actors may

202

participate in movies with different ratings but ratings of movies

203

by the same directors are more similar. A single feature can also

204

be made more precise by taking advantage of mutual information

205

[12], e.g. using frequent words or information retrieval methods

206

like TF (term frequency), DF (document frequency), and TF-IDF

207

(term frequency-inverse document frequency). This was employed

208

by the hybrid recommendation system of Rubtsov et al. [14] that

209

used the top-2000 most frequent words in titles as a feature in their

210

recommendation system opposed to all the words.

211

4

METHODOLOGY

212

This section describes the methodology employed for answering

213

the research questions. First, the hybrid recommendation model.

214

This is followed by a description of the data that is provided to the

215

recommendation systems. Furthermore, the metrics for evaluating

216

the performance of both recommendation systems are presented

217

and the experimental setup is described.

218

4.1

The Hybrid Recommendation Model

219

The hybrid recommendation model uses a feature combination

tech-220

nique and consists of a matrix factorisation model that incorporates

221

item information. The model is implemented using the LightFM

222

library [8], which is a Python implementation of a matrix

factori-223

sation model that can deal with user and item information. This

224

model acts as a standard matrix factorisation model when no user

225

or item information is provided.

226

The LightFM model represents each user and/or item as a

com-227

bination of latent representations. For example, the representation

228

of the series ‘Beste Zangers‘ is a combination of the

representa-229

tion of the genre music, the genre amusement and the broadcaster

230

AVROTROS. The latent representation approach is utilised in the

231

hybrid model for each item, so if the genre music and the genre

232

amusement are both liked by the same users then their embeddings

233

will be close together; if both genres are never liked by the same

234

users then their embeddings will be far apart. The dimensionality

235

of these latent feature embeddings can be optionally adjusted in

236

the model.

237

The LightFM library offers two types of loss functions for implicit

238

feedback learning-to-rank: the WARP (Weighted

Approximate-239

Rank Pairwise) loss [17] and BPR (Bayesian Personalised Ranking)

240

loss. The LightFM documentation states that the WARP loss

typi-241

cally performs better than BPR so this function has been chosen

242

for the implementation of the hybrid recommendation model. The

243

WARP loss samples a negative item for each (user, positive item)

244

pair and computes predictions for both positive and negative items.

245

A gradient update is performed if the prediction of the negative item

246

is valued higher than that of the positive item, otherwise negative

247

items are continuously sampled until a higher negative prediction

248

does occur.

249

The execution of the LightFM model can be sped up by making

250

use of the offered multi-threading during the training, prediction

251

and evaluation of the model [9]. This can however lead to a decrease

252

in accuracy when the interaction matrix is dense, but does not lead

253

to a measurable loss of accuracy when a sparse data set is trained.

254

4.2

The Data

255

The input data for the recommendation systems consists of

interac-256

tion information and content features.

257

4.2.1 Interaction Information. The first data set consists of

inter-258

action information that is provided by the event data of the NPO.

259

The event data describes all interactions that users have had with

260

the NPO Start service, e.g. clicks, stream starts and refreshes. The

261

interaction information spans a period of 22 days of which the first

262

21 days are intended for training the model and the last day for

263

testing. A period of 21 days was used since that is the sliding

win-264

dow used by the current NPO Start recommendation system (see

265

section 2.1). The event data of the period 1 to 22 March, 2019 was

266

used as interaction information for the NPO Start recommendation

267

system and the collaborative filtering part of the hybrid

recommen-268

dation system. This event data was pre-processed to only gather

269

interaction information about the watched series of users, on the

270

condition that one episode of a series needs to be watched for at

271

least half its duration in order to be included. The total interaction

272

information consists of 1,235,728 interactions and the distribution

273

of these interactions is shown in Figure 1. A weekly pattern of

274

interactions is visible in the distribution and the average amount

275

of interactions per day is about 56,000.

276

Figure 1: Distribution of the Amount of Interactions per Day A total of 181,357 users were identified and on average about

277

42,000 unique users had at least one interaction per day. The

dis-278

tribution of the amount of unique series each user interacted with

279

is shown in 2a. This distribution is skewed to the right, indicating

280

that the majority of users has only watched a couple of series in

281

this period. However, there is a long tail of a few users who have

282

had tens of interactions.

283

(5)

A total of 1446 unique series were watched and on average about

284

808 series were watched per day in this period. The distribution

285

of the amount of unique users that watched a series is shown in

286

2b. This distribution is also skewed to the right, indicating that the

287

majority of series was watched by a couple to 5000 users. A big

288

portion was watched by 5000 to 10,000 users. The distribution has

289

got a long tail, indicating that there are a couple of series which

290

were watched by a great number of users, thus were well-rated.

291

The tail does contain a few spikes of series that were watched by

292

roughly the same amount of users.

293

(a) Amount of Unique Series Interactions per User

(b) Amount of Unique User Inter-actions per Series

Figure 2: Distribution of the Amount of Interactions for Users and Series

4.2.2 Content Features. The Publieke Omroep Media Service (POMS)

294

contains information about all content that is offered by the NPO,

295

which ranges from broadcasts and movies to podcasts. This amounts

296

to a total of about 1.5 million media items. Each item consists of

297

37 columns describing metadata, e.g. a media id (mid), age rating,

298

broadcaster, credits, descriptions, genres, images, etc. These items

299

were pre-processed to only gain broadcasts which are available to

300

stream on the NPO Start service, resulting in about 84,000

broad-301

casts. Each broadcast has its own media id and a series reference

302

which refers to the series this broadcast is a part of. A total of 2490

303

series were identified and a series can consist of a lot of individual

304

broadcasts, a couple broadcasts or a single broadcast. The series

305

NOS Journaal is an example of a series that consists of a lot of

306

individual broadcasts since it has 10,297 broadcasts. Out of all the

307

metadata, six metadata features were selected to be used in the

308

content-based part of the recommendation system, namely

broad-309

caster, credits, description, genres, subtitles and title. The content

310

features broadcaster, credits, description, genres and title were

se-311

lected because they are commonly used as features in content-based

312

recommendation systems (see section 3.3). The feature subtitles is

313

often not provided in multimedia content and has been included

314

since it gives more in-depth information about the content in a

315

broadcast than the title or description. The content features are

316

described in Appendix A Table 6. The POMS data is grouped per

se-317

ries and aggregated based on unique values per item. All metadata

318

is provided by program makers and can differ in completeness and

319

detail. The percentage series with missing values for the content

320

features is displayed in Figure 3 to investigate the completeness.

321

This shows that the features broadcaster and title are complete

322

for all series, however some are missing information about its

de-323

scription and genres. Furthermore, 40% of the series do not have

324

information about credits and subtitles, which amounts to about

325

1000 series.

Figure 3: Percentage Series with Missing Values for the Con-tent Features

326

Half of the six content features consist of categorical features and

327

the other half are textual features. The three content features

broad-328

caster, credits and genres are categorical. The remaining features

329

title, description and subtitles are textual.

330

Categorical features. A total of 30 unique broadcasters were

iden-331

tified for the series and each series has at least one or several

332

broadcasters associated with it. The percentage of how often a

333

broadcaster is associated with a series is shown in Figure 4. Most

334

series are broadcasted by the VPRO and other frequent broadcasters

335

are the NTR, AVTR (AVROTROS) and BNVA (BNNVARA).

Figure 4: Percentage Series with Broadcaster

336

5383 unique credits were identified and a series either has

mul-337

tiple credits, one credit or no credits associated with it. A single

338

person is often accredited in a single series and sometimes in several.

339

However, there are a few people, like Sophie Hilbrand, Tom Egbers

340

and Astrid Kersseboom, that are accredited more than ten times. It

341

should be noted that the credits are dirty, since a big portion of the

342

series does not include this feature and the amount of accredited

343

people per series can range from hundreds of people to only one

344

person.

345

Series either had multiple genres, one genre or no genre assigned

346

to them. A total of 53 different genres were identified and the

347

percentage of how often a genre is associated with a series is shown

348

in Appendix A Figure 10. Genres have a main type indicated by an

349

id of four integers, e.g. 3.0.1.7 ‘Informatief‘, and may have sub-types

350

(6)

which are indicated by an identifier of five integers, e.g. 3.0.1.6.26

351

‘Informatief, Religieus‘. About 30% of the series have the genre

352

‘Informatief‘ which amounts to about 750 series. Other frequent

353

genres for series are ‘Amusement‘, ‘Jeugd‘ en ‘Documentaire‘, which

354

each occur about 13%.

355

Textual features. Additionally, there are three textual features,

356

namely title, description and subtitles. As mentioned before, all

357

content features are grouped per series and all unique values are

358

aggregated. This means that all unique broadcast values for the

359

textual features were concatenated for each series. The average

360

word count of these features and their median per series is shown

361

in Table 1. It should be noted that the word count only includes

362

series that have data for that particular feature. The feature title has

363

the lowest mean word count, followed by description and subtitles

364

has the highest. All the medians of the features lie below the mean,

365

indicating that the length of that particular feature is not evenly

366

distributed. The distribution of the title, description and subtitles

367

word count is displayed in Figure 5 on a logarithmic x-scale. The

368

three distributions all indicate a skewed distribution to the right,

369

meaning that a big portion of the series has textual features with a

370

low word count. The long tail of the distributions indicate that a

371

few series have high word counts for the textual features.

372

Table 1: Mean and Median Word Count for the Textual Fea-tures

Feature Mean Median

Title 8.2 4.0

Description 391.9 76.0

Subtitles 58951.1 12434.0

(a) Title (b) Description

(c) Subtitles Figure 5: Distributions of Word Count for the Textual Fea-tures

4.3

Feature Encoding

373

The interaction information and content features were prepared

374

into the right format before being provided to the hybrid

recom-375

mendation model.

376

4.3.1 Interaction information. The interaction information was

377

processed into (user, series) pairs and transformed into a user

inter-378

action matrix.

379

4.3.2 Content features. For the textual features, some language

pro-380

cessing was employed during pre-processing. This pre-processing

381

consisted of lowercasing and tokenising the words for the features.

382

Afterwards, the punctuation, tokens smaller than four letters and

383

dutch & english stopwords were removed. Finally, TF-IDF was

per-384

formed where each series was regarded as a text and the whole set

385

of series as a document. Feature selection was employed to improve

386

the quality of these content features. For the title, the three words

387

with the highest TF-IDF were extracted per series, the top ten for

388

the description and the top 20 for the subtitles.

389

Afterwards, all content features were exploded into (series,

fea-390

ture value) pairs and one-hot-encoded using scikit-learn’s

DictVec-391

torizer class [5]. This produced a dictionary for each series, where

392

the key is the feature value and the weights are the values, and

393

was transformed into an item information matrix. The amount of

394

unique feature values for each content feature is displayed in Table

395

2.

396

Table 2: Amount of Unique Feature Values per Content Fea-ture

Content Features Amount of Unique Feature Values

Broadcaster 28 Genres 53 Credits 5689 Title 3252 Description 13396 Subtitles 14210

4.4

Evaluation

397

The performance of a recommendation system is assessed by the

398

quality of recommendations. The quality was evaluated by the two

399

metrics mean precision@k (mean p@k) and mean reciprocal rank

400

(MRR).

401

4.4.1 Mean Precision@k. Mean precision@k is a metric that

eval-402

uates the average proportion of top-k recommended items that are

403

relevant to users. A relevant item is an item that was chosen by a

404

user when it was offered in a ribbon. Relevant items are denoted as

405

a true positive (TP) which are positive predicted values. The

preci-406

sion is thus denoted as the total number of predicted positives out

407

of all predicted items. The equation for the precision@k is shown

408

in equation 1.

409

P@k = |{i ∈ T P | i ranked in top k}|k (1)

The precision@k is evaluated over all recommendations and

410

averaged into the mean precision@k per user to evaluate the overall

411

quality of the system (see equation 2).

412

MeanP@k =

ÍN

n=1P@k(n)

N (2)

4.4.2 Mean Reciprocal Rank. Mean reciprocal rank is a metric that

413

evaluates the average ranking quality of recommendation lists that

414

a model produces. This metric evaluates how successful the model

415

is in ranking the highest relevant item to users, so it measures

416

how many non relevant recommendations users have to skip in

417

(7)

their ranked list until the first relevant recommendation. The mean

418

reciprocal rank is calculated by dividing the best possible rank

419

by the actual rank of the first relevant item and averaging it (see

420 equation 3). 421 MRR =N1 N Õ i=1 1 ranki (3)

The higher the value of the performance metrics, the better. The

422

version with the highest mean precision has the most success of

423

recommending items that users are interested in, and the version

424

with the highest MRR is most successful in ranking the highest

425

relevant item in a personalised manner.

426

4.5

The Experimental Setup

427

The experimental setup consists of three parts that each correspond

428

to a research question.

429

Interaction information was split into a train and test set for

430

the experimental setup. As mentioned in section 4.2.1, a period of

431

21 days was used for the train set and the following day after the

432

train period was used as test set for the recommendation systems.

433

The train set consists of a total of 1,192,556 interactions and the

434

test set of 41,538 interactions. The interactions of the train set

435

were performed by 179,714 unique users and the test set by 31,127

436

users. An additional test set was constructed from the original

437

test set, called the “recommended test set“, which consists of the

438

interactions that occurred on the top-k series that were actually

439

recommended in the ‘Aanbevolen voor jou‘ ribbon on the NPO

440

Start service. The recommended test set was included since this is

441

the subset of interactions whereof information is available about

442

the top-k precision and rank of these items. Ak of 5 was used for

443

the recommended test set, since that is the typical amount of items

444

that is visible on a ribbon of the NPO Start service (see section 2.1).

445

The test set ended up with 149 interactions that were performed by

446

124 users. It should be noted that interactions of users that were

447

present in the test sets but not in the train set were removed from

448

the experiment. The sparsity of the used interaction information is

449

0.22%, thus this data is sparse.

450

4.5.1 RQ1. The experimental setup of the NPO Start

recommenda-451

tion system is shown in Figure 6a. It starts with supplying the train

452

set to the NPO Start model as described in section 2.1. Afterwards,

453

the predictions of this model are evaluated against the test sets

454

using the performance metrics.

455

4.5.2 RQ2. The experimental setup of the hybrid

recommenda-456

tion system is shown in Figure 6b. The same train and test sets are

457

used in this system as the NPO Start recommendation system. The

458

interaction matrix of the train set and the feature matrix of the

con-459

tent features was supplied to the hybrid recommendation model

460

as described in section 4.1. The model used these two matrices

461

for training and serving out ranked item predictions for each user

462

present in the test sets. Lastly, the performance of the hybrid

recom-463

mendation system was evaluated against the test sets of interaction

464

information using the performance metrics. Multi-threading was

465

used during the training, prediction and evaluation of the model to

466

speed up the execution and it should not lead to a measurable loss

467

of accuracy in this case since the interaction information is sparse

468

(see section 4.1).

469

The described experimental setup was used for performing 64

470

different experiments on the hybrid recommendation model. Each

471

experiment used the same interaction information and a different

472

set of content features while training the hybrid model. The first

473

experiment acted as a baseline wherein no content features were

474

supplied to the model and the other 63 different experiments

exper-475

iments used a different combination of the six content features (see

476

Table 6 for all combinations). The combinations start with a single

477

content feature, go to combinations of two content features and

478

end with a combination that incorporates all content features. For

479

example, experiment 16 incorporates the description and genres

480

features into the hybrid model. Each experiment model was trained

481

on a range from 0 to 100 epochs with a step size of 10 on standard

482

settings and its predictions were evaluated against both test sets

483

to investigate the learning curve of each model. The experiment

484

model that accomplished the highest performance was compared

485

to the baseline model afterwards.

486

4.5.3 RQ3. The last part of the experimental setup compared the

487

performance of the NPO Start recommendation system to that of

488

the hybrid recommendation system.

489

The current recommendation system used the same experimental

490

setup described above in section 4.5.1.

491

The hybrid recommendation system used the experimental setup

492

for the experiment model that accomplished the highest

perfor-493

mance as described above in 4.5.2. The hyperparameters of this

494

model were optimised using a tree based regression model from

495

the scikit-optimize library [5], which allows for finding the optimal

496

hyperparameters to maximise model performance. The optimal

497

hyperparameters were then used for producing recommendations

498

with this model.

499

Afterwards, the performance metrics of both systems were

com-500

pared to one another.

501

5

RESULTS

502

5.1

RQ1: What is the Performance of the NPO

503

Start Recommendation System?

504

The NPO Start recommendation system offers recommendations to

505

a user several times a day (see section 2.1) and since the performance

506

metrics are evaluated per user, the precision@5 and reciprocal rank

507

results of an offer were first averaged per user before producing

508

the final results. Since there was no precision@5 and reciprocal

509

rank information available for all the interaction information of

510

the full test set, all user interactions who were not included in the

511

recommended test set were rewarded a precision@5 and reciprocal

512

rank of 0. The results of the performance metrics for both test sets

513

are summarised in Table 3.

514

Table 3: Results of the NPO Start Recommendation System

Metric Mean Std

Mean p@5 0.00 0.01

MRR 0.00 0.04

(a) Test Set

Metric Mean Std

Mean p@5 0.19 0.06

MRR 0.61 0.34

(b) Recommended Test Set

(8)

(a) The NPO Start Recommendation System

(b) The Hybrid Recommendation System Figure 6: The Experimental Setup Mean Precision@5. The mean p@5 of 0.00 and standard deviation

515

of 0.01 on the test set indicates that there were almost no relevant

516

items for a user in the recommendations provided by the NPO Start

517

recommendation system. This indicates that hardly any series that

518

were recommended to a user were found being relevant.

519

However, the mean p@5 of 0.19 on the recommended test set

520

indicates that on average 1 in 5 recommendations is a relevant

521

item for a user. This indicates that on average a user needs to see a

522

ranked list of 5 items once in order to find a series that suits him

523

or her. The boxplot of this metric (see Figure 7a) indicates that a

524

majority of users have a mean p@5 of 0.20 and there are a few

525

occasions where on average 2 in 5 recommendations is relevant or

526

about 1 in 10 recommendations is relevant.

527

Mean Reciprocal Rank. The mean p@5 of 0.00 and standard

de-528

viation of 0.04 on the test set indicates that there were almost no

529

relevant items for a user in the recommendations, since a reciprocal

530

rank of 0 indicates that there was no highest relevant item in the

531

offered recommendations. This illustrates that hardly any series

532

that were recommended fit the user.

533

Since a reciprocal rank of 1.0 indicates that the highest relevant

534

item was the first item and a reciprocal rank of 0.5 indicates it being

535

being the second item, a MRR of 0.61 on the recommended test

536

set indicates that on average the highest relevant item is placed as

537

the second item in a ranked list for a user. The MRR does have a

538

high standard deviation and the boxplot (see Figure 7b) indicates

539

that 50% of users their MRR ranges from a score of 1.0 to 0.33, thus

540

being ranked as either the first, second or third item.

541

(a) Mean P@5 (b) MRR

Figure 7: Boxplots of the NPO Start Recommendation Sys-tem Results per User

5.2

RQ2: Which Content Features Improve the

542

Performance of the Hybrid

543

Recommendation System the Most?

544

The results of the 64 different experiment models of the hybrid

545

recommendation system for the test set and the recommended test

546

set is shown in Appendix B Figure 11 & 12.

547

The results of the experiment models on the test set, which

548

contains all watched interactions, show smooth learning curves for

549

both performance metrics. The learning curves show an exponential

550

rise with increasing epochs.

551

The results of the experiment models on the recommended test

552

set, which contained all watched interactions that were

recom-553

mended, show learning curves that are erratic. However, the

learn-554

ing curves do show an increase in metrics with increasing epochs.

555

Mean Precision@5. The top-10 mean p@5 of content feature

556

combinations is shown in Table 8. The mean p@5 and standard

557

deviation results for the top-10 combinations of each test set are

558

(9)

very similar. Overall, the precision of the test set results are slightly

559

better than that of the recommended test set.

560

The experiment model that incorporated the 29th combination of

561

content features accomplished the highest mean p@5 for the test set.

562

This combination used the content features broadcaster, genres and

563

title. The results of this experiment model and that of the baseline

564

model, which uses no content features, is shown in Figure 8a. A two

565

sample right-tailed z-test (α = 0.05) was conducted to compare the

566

mean p@5 of experiment model 29 and the baseline model. There

567

was no significant difference in the scores for experiment model

568

29 ( ¯x = 0.13, s = 0.12, n = 31127) and the baseline model (¯x = 0.13,

569

s = 0.12, n = 31127); z = 0.09, p = 0.46. These results suggest

570

that the best experiment model that used content features does not

571

perform better than the baseline model based on the mean p@5 of

572

the test set.

573

The experiment model that incorporated the 48th combination

574

of content features accomplished the highest mean p@5 for the

575

recommended test set. This combination used the content features

576

broadcaster, description, genres and title. The results of this

experi-577

ment model and that of the baseline model, which uses no content

578

features, is shown in Figure 8b. A two sample right-tailed z-test

579

(α = 0.05) was conducted to compare the mean p@5 of experiment

580

model 48 and the baseline model. There was no significant

differ-581

ence in the scores for experiment model 48 ( ¯x = 0.11, s = 0.10,

582

n = 124) and the baseline model (¯x = 0.09, s = 0.10, n = 124);

583

z = 1.36, p = 0.09. These results suggest that the best experiment

584

model that used content features does not perform better than the

585

baseline model based on the mean p@5 of the recommended test

586

set.

587

(a) Test Set (b) Recommended Test Set

Figure 8: Mean P@5 Results of the Baseline and Top Combi-nation

The experiment model that achieved the highest mean p@5

588

on both tests used similar content features. Experiment model 48

589

included the description feature which was not used in experiment

590

model 29. The results of these experiment models indicate that

591

incorporating content features into the hybrid recommendation

592

model does not necessarily improve the mean p@5.

593

Mean Reciprocal Rank. The top-10 MRR of content feature

com-594

binations is shown in Table 9. The MRR and standard deviation

595

results for the top-10 combinations of each test set are very similar.

596

Overall, the precision of the test set results are a little higher than

597

that of the recommended test set

598

The experiment model that incorporated the 48th combination

599

of content features accomplished the highest MRR for the test set.

600

This combination used the content features broadcaster, description,

601

genres and title. The results of this experiment model and that of the

602

baseline model, which uses no content features, is shown in Figure

603

9a. A two sample right-tailed z-test (α = 0.05) was conducted to

604

compare the MRR of experiment model 48 and the baseline model.

605

There was no significant difference in the scores for experiment

606

model 48 ( ¯x = 0.37, s = 0.35, n = 31127) and the baseline model

607

( ¯x = 0.37, s = 0.35, n = 31127); z = 0.06, p = 0.48. These results

608

suggest that the best experiment model that used content features

609

does not perform better than the baseline model based on the MRR

610

of the test set.

611

The experiment model that incorporated the 48th combination

612

of content features accomplished the highest MRR for the

rec-613

ommended test set. This combination used the content features

614

broadcaster, description, genres and title. The results of this

experi-615

ment model and that of the baseline model, which uses no content

616

features, is shown in Figure 9b. A two sample right-tailed z-test

617

(α = 0.05) was conducted to compare the MRR of experiment model

618

48 and the baseline model. There was no significant difference in

619

the scores for experiment model 48 ( ¯x = 0.31, s = 0.32, n = 124)

620

and the baseline model ( ¯x = 0.28, s = 0.31, n = 124); z = 0.76,

621

p = 0.22. These results suggest that the best experiment model that

622

used content features does not perform better than the baseline

623

model based on the MRR of the recommended test set.

624

(a) Test Set (b) Recommended Test Set

Figure 9: MRR Results of the Baseline and Top Combination Experiment model 48 achieved the highest MRR on both tests.

625

The results of this experiment model indicates that incorporating

626

content features into the hybrid recommendation model does not

627

necessarily improve the MRR.

628

5.3

RQ3: Can the Performance of the NPO Start

629

Recommendation System be Improved by

630

Implementing a Hybrid Recommendation

631

System?

632

The majority of the results of the previous section indicate that

633

the experiment model that incorporated the 48th combination

per-634

formed similar but slightly better than the baseline, so the

hyper-635

parameters of the model that used the content features broadcaster,

636

description, genres and title were optimised. The resulting

hyper-637

parameters after the optimisation is shown in Table 4 and the

ac-638

companying results for the performance metrics is shown in Table

639

5.

640

Mean Precision@5. Two sample right-tailed z-tests (α = 0.05)

641

were conducted to compare the mean p@5 results of the optimised

642

model to the NPO Start model.

643

When evaluated on the test set, there was a significant difference

644

in the scores for the optimised model ( ¯x = 0.19, s = 0.11) and the

645

NPO Start model ( ¯x = 0.00, s = 0.01); z = 296.8, p = 0.00. These

646

results suggest that a recommendation system using the optimised

647

(10)

Table 4: Hyperparameter Values of the Optimised Model Hyperparameter Value Epochs 89 Learning rate 0.01 Number of components 168 Item alpha 0.00 Scaling 0.06

Table 5: Results of the Optimised 48th Experiment Model

Metric Mean Std

Mean p@5 0.19 0.11

MRR 0.53 0.34

(a) Test Set

Metric Mean Std

Mean p@5 0.13 0.10

MRR 0.39 0.32

(b) Recommended Test Set

model performs better than a recommendation system using the

648

NPO Start model based on their mean p@5 of the test set.

649

When evaluated on the recommended test set, there was no

650

significant difference in the scores for the optimised model ( ¯x =

651

0.13,s = 0.10) and the NPO Start model (¯x = 0.19, s = 0.06);

652

z = −5.17, p = 1.00. These results suggest that a recommendation

653

system using the optimised model does not perform better than a

654

recommendation system using the NPO Start model based on their

655

mean p@5 of the recommended test set.

656

The results of both test sets indicate that the performance of

657

the hybrid recommendation system in comparison to the NPO

658

Start recommendation system based on the metric mean p@5 is

659

dependent on the set of interaction information used for evaluation.

660

Mean Reciprocal Rank. Two sample right-tailed z-tests (α = 0.05)

661

were also conducted to compare the MRR results of the optimised

662

model to the NPO Start model.

663

When evaluated on the test set, there was a significant difference

664

in the scores for the optimised model ( ¯x = 0.53, s = 0.34) and the

665

NPO Start model ( ¯x = 0.00, s = 0.04); z = 268.5, p = 0.00. These

666

results suggest that a recommendation system using the optimised

667

model performs better than a recommendation system using the

668

NPO Start model based on their MRR of the test set.

669

When evaluated on the recommended test set, there was no

670

significant difference in the scores for the optimised model ( ¯x =

671

0.39,s = 0.32) and the NPO Start model (¯x = 0.61, s = 0.34);

672

z = −4.48, p = 1.00. These results suggest that a recommendation

673

system using the optimised model does not perform better than a

674

recommendation system using the NPO Start model based on their

675

MRR of the recommended test set.

676

The results of both test sets indicate that the performance of the

677

hybrid recommendation system in comparison to the NPO Start

678

recommendation system based on the metric MRR is dependent on

679

the set of interaction information used for evaluation.

680

6

CONCLUSIONS

681

In this thesis, a hybrid recommendation system that utilises

meta-682

data was presented and compared to the current recommendation

683

system of the NPO Start service which uses collaborative filtering.

684

The hybrid recommendation system serves out predictions using a

685

hybrid LightFM model to which interaction information and

con-686

tent features are supplied. The content features consist of the six

687

metadata features broadcaster, credits, description, genres, subtitles

688

and title.

689

Based on experiments where different combinations of the

con-690

tent features were supplied to the hybrid model, the results

in-691

dicated that the model that utilised the broadcaster, description,

692

genres and title features performed similar but slightly higher than

693

the model that utilised no content features. This concludes that

694

incorporating content features into the hybrid recommendation

695

model does not necessarily improve the performance.

696

Based on the comparison of the optimised best performing hybrid

697

recommendation model and the current model of the NPO start

698

recommendation system, the results indicated that the performance

699

of the hybrid recommendation system is better than that of the NPO

700

Start recommendation system when based on a broader evaluation

701

set. This concludes that a hybrid recommendation using metadata

702

can perform better than the current recommendation system of

703

NPO Start.

704

7

DISCUSSION

705

This section discusses the results and the limitations of the

em-706

ployed methodology. Also, possible future work is presented that

707

could be taken to overcome these limitations.

708

The results indicated that the hybrid recommendation model

709

does not perform better when content features are used opposed to

710

when no content features are used. This does not fit with previous

711

research stating that incorporating content into a collaborative

712

filtering approach provides more accurate recommendations when

713

ratings are sparse [1, 8]. One possible cause for this result is the

714

completeness of the used metadata for the content features.

Pre-715

vious research has shown that features derived from high-quality

716

metadata lead to a better performance of content-based

recommen-717

dation systems [8, 14, 15]. As mentioned in section 4.2.2, the used

718

metadata differed in completeness, detail and had missing values

719

for a portion of the content features. This indicates that a better

720

performance could have been achieved in the hybrid

recommen-721

dation experiment models that used content features when the

722

metadata was of a better quality. Further research is needed to

es-723

tablish if the quality of the metadata was a limitation of evaluating

724

the performance of content features in the hybrid recommendation

725

model.

726

Furthermore, the results demonstrated that the hybrid

recom-727

mendation system performs significantly better than the NPO Start

728

recommendation system when evaluated on the full test set opposed

729

to the recommended test set. The recommended test set consists

730

of watched series that were recommended to users and assumes

731

that users would have interacted with the same series, regardless

732

of which model was used to generate the recommendations. This

733

assumption is a major drawback of offline experiments [4], since the

734

recommended test set does not take into account how different the

735

hybrid recommendation model is compared to the NPO Start model.

736

The NPO Start recommendation system recommends well-rated

737

series to users, opposed to the hybrid recommendation system that

738

recommends a mix of well-rated and similar series to users. This

739

(11)

is apparent in the huge loss of performance when the NPO Start

740

recommendation system was evaluated on the full test set instead

741

of the recommended test set. The full test set gives a broader view

742

of relevant items for users and is thus more suited when

compar-743

ing the two recommendation systems to each other. However, the

744

most reliable performance results are achieved when both

recom-745

mendation systems are compared in an online setting, because this

746

evaluates the recommendations on actual user behaviour.

747

Lastly, the generalisability of the results is limited by the

inter-748

action information used in the experimental set-up, since only one

749

specific time period was used. Different performance results could

750

be achieved for the recommendation systems in different time

peri-751

ods because of the temporality of interactions. A more collaborative

752

approach could be favoured in one time period based on a higher

753

occurrence of well-rated series, e.g. the series ‘Poldark‘ generated a

754

high amount of interactions in a short time because it was heavily

755

promoted inside the NPO Start service and on social media.

Alter-756

natively, a more content-based approach could be favoured based

757

on events happening in the world, e.g. users watch more content

758

about the Dutch royal families close to Kingsday. Future research

759

is needed to establish the performance of the recommendation

760

systems in different time periods and how temporality could be

761

incorporated to improve the hybrid recommendation model.

762

ACKNOWLEDGMENTS

763

I would like to thank the Marketing Intelligence Team at the NPO

764

for entrusting me with this project and for providing a welcoming

765

and supportive environment to work in. I would especially like to

766

thank Robbert van Waardhuizen for supervising me internally in

767

the company. Additionally, I am grateful for the helpful observations

768

provided by Dr Maarten Marx.

769

REFERENCES

[1] Marie Al-Ghossein, Pierre-Alexandre Murena, Talel Abdessalem, Anthony Barré, and Antoine Cornuéjols. 2018. Adaptive collaborative topic modeling for online recommendation. Proceedings of the 12th ACM Conference on Recommender Systems (2018), 338–346.

[2] Robert Bell, Yehuda Koren, and Chris Volinsky. 2007. Modeling relationships at multiple scales to improve accuracy of large recommender systems. In Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 95–104.

[3] Robin Burke. 2002. Hybrid recommender systems: Survey and experiments. User modeling and user-adapted interaction 12, 4 (2002), 331–370.

[4] Carlos A Gomez-Uribe and Neil Hunt. 2016. The netflix recommender system: Algorithms, business value, and innovation. ACM Transactions on Management Information Systems (TMIS) 6, 4 (2016), 13.

[5] Tim Head, MechCoder, Gilles Louppe, Iaroslav Shcherbatyi, fcharras, Zé Viní-cius, cmmalone, Christopher Schröder, nel215, Nuno Campos, Todd Young, Ste-fano Cereda, Thomas Fan, rene rex, Kejia (KJ) Shi, Justus Schwabedal, carlos-danielcsantos, Hvass-Labs, Mikhail Pak, SoManyUsernamesTaken, Fred Call-away, Loïc Estève, Lilian Besson, Mehdi Cherti, Karlson Pfannschmidt, Fabian Linzberger, Christophe Cauet, Anna Gut, Andreas Mueller, and Alexander Fabisch. 2018. scikit-optimize/scikit-optimize: v0.5.2. (March 2018). https: //doi.org/10.5281/zenodo.1207017

[6] Yehuda Koren. 2009. The bellkor solution to the netflix grand prize. Netflix prize documentation 81, 2009 (2009), 1–10.

[7] Yehuda Koren, Robert Bell, and Chris Volinsky. 2009. Matrix factorization tech-niques for recommender systems. Computer 8 (2009), 30–37.

[8] Maciej Kula. 2015. Metadata Embeddings for User and Item Cold-start Recom-mendations. In Proceedings of the 2nd Workshop on New Trends on Content-Based Recommender Systems co-located with 9th ACM Conference on Recommender Sys-tems (RecSys 2015), Vienna, Austria, September 16-20, 2015. (CEUR Workshop Pro-ceedings), Toine Bogers and Marijn Koolen (Eds.), Vol. 1448. CEUR-WS.org, 14–21. http://ceur-ws.org/Vol-1448/paper4.pdf

[9] Maciej Kula. 2016. Welcome to LightFM’s documentation! (2016). https://lyst. github.io/lightfm/docs/index.html

[10] Pasquale Lops, Marco De Gemmis, and Giovanni Semeraro. 2011. Content-based recommender systems: State of the art and trends. In Recommender systems handbook. Springer, 73–105.

[11] Malte Ludewig, Iman Kamehkhosh, Nick Landia, and Dietmar Jannach. 2018. Effective Nearest-Neighbor Music Recommendations. In Proceedings of the ACM Recommender Systems Challenge 2018. ACM, 3.

[12] Christopher Manning, Prabhakar Raghavan, and Hinrich Schütze. 2010. Intro-duction to information retrieval. Natural Language Engineering 16, 1 (2010), 100–103.

[13] Michael J Pazzani. 1999. A framework for collaborative, content-based and demographic filtering. Artificial intelligence review 13, 5-6 (1999), 393–408. [14] Vasiliy Rubtsov, Mikhail Kamenshchikov, Ilya Valyaev, Vasiliy Leksin, and

Dmitry I Ignatov. 2018. A hybrid two-stage recommender system for auto-matic playlist continuation. In Proceedings of the ACM Recommender Systems Challenge 2018. ACM, 16.

[15] Márcio Soares and Paula Viana. 2015. Tuning metadata for better movie content-based recommendation systems. Multimedia Tools and Applications 74, 17 (2015), 7015–7036.

[16] Maksims Volkovs, Himanshu Rai, Zhaoyue Cheng, Ga Wu, Yichao Lu, and Scott Sanner. 2018. Two-stage model for automatic playlist continuation at scale. In Proceedings of the ACM Recommender Systems Challenge 2018. ACM, 9. [17] Jason Weston, Samy Bengio, and Nicolas Usunier. 2011. Wsabie: Scaling up to

large vocabulary image annotation. In Twenty-Second International Joint Confer-ence on Artificial IntelligConfer-ence.

[18] Hojin Yang, Yoonki Jeong, Minjin Choi, and Jongwuk Lee. 2018. Mmcf: Multi-modal collaborative filtering for automatic playlist continuation. In Proceedings of the ACM Recommender Systems Challenge 2018. ACM, 11.

[19] Hamed Zamani, Markus Schedl, Paul Lamere, and Ching-Wei Chen. 2018. An Analysis of Approaches Taken in the ACM RecSys Challenge 2018 for Auto-matic Music Playlist Continuation. Proceedings of the 12th ACM Conference on Recommender Systems (2018), 527–528.

(12)

A

THE DATA

Table 6: Overview of the Content Features

Feature Type Description

Broadcaster Categorical Broadcaster of the broadcast, e.g. NOS.

credits list The people accredited in the broadcast,

such as presenters or guests. Description String Description of the broadcast. This is

ei-ther the main description, oei-therwise the short description or the kicker.

Genres List Genres of the broadcast denoted by a

genre id and name, e.g. (3.0.1.6, [Amuse-ment]).

Subtitles String The subtitles of the broadcast, which were extracted using the POMS subti-tles API.

Title String The main title of the broadcast.

Figure 10: Percentage Series with Genres

Table 7: The Content Features Combinations

Index Feature 0 None 1 Broadcaster 2 Credits 3 Description 4 Genres 5 Title 6 Subtitles 7 Broadcaster, credits 8 Broadcaster, description 9 Broadcaster, genres 10 Broadcaster, title 11 Broadcaster, subtitles 12 Credits, description 13 Credits, genres 14 Credits, title 15 Credits, subtitles 16 Description, genres 17 Description, title 18 Description, subtitles 19 Genres, title 20 Genres, subtitles 21 Title, subtitles

22 Broadcaster, credits, description

23 Broadcaster, credits, genres

24 Broadcaster, credits, title

25 Broadcaster, credits, subtitles

26 Broadcaster, description, genres

27 Broadcaster, description, title

28 Broadcaster, description, subtitles

29 Broadcaster, genres, title

30 Broadcaster, genres, subtitles

31 Broadcaster, title, subtitles

32 Credits, description, genres

33 Credits, description, title

34 Credits, description, subtitles

35 Credits, genres, title

36 Credits, genres, subtitles

37 Credits, title, subtitles

(13)

38 Description, genres, title

39 Description, genres, subtitles

40 Description, title, subtitles

41 Genres, title, subtitles

42 Broadcaster, credits, description, genres

43 Broadcaster, credits, description, title

44 Broadcaster, credits, description, subtitles

45 Broadcaster, credits, genres, title

46 Broadcaster, credits, genres, subtitles

47 Broadcaster, credits, title, subtitles

48 Broadcaster, description, genres, title

49 Broadcaster, description, genres, subtitles

50 Broadcaster, description, title, subtitles

51 Broadcaster, genres, title, subtitles

52 Credits, description, genres, title

53 Credits, description, genres, subtitles

54 Credits, description, title, subtitles

55 Credits, genres, title, subtitles

56 Description, genres, title, subtitles

57 Broadcaster, credits, description, genres, title

58 Broadcaster, credits, description, genres, subtitles

59 Broadcaster, credits, description, title, subtitles

60 Broadcaster, credits, genres, title, subtitles

61 Broadcaster, description, genres, title, subtitles

62 Credits, description, genres, title, subtitles

63 Broadcaster, credits, description, genres, title, subtitles

B

RESULTS

Figure 11: Results for the Content Feature Combinations on the Test Set

Figure 12: Results for the Content Feature Combinations on the Recommended Test Set

(14)

Table 8: The Top-10 Mean P@5 Content Feature Combina-tion Results

Rank Combination Epoch Mean p@5 Std

1 29 100 0.13 0.12 2 0 100 0.13 0.12 3 48 100 0.13 0.12 4 38 100 0.13 0.12 5 8 100 0.13 0.12 6 16 100 0.13 0.12 7 19 100 0.13 0.12 8 26 100 0.13 0.12 9 17 100 0.12 0.12 10 3 100 0.12 0.12

(a) Test Set

Rank Combination Epoch Mean p@5 Std

1 48 100 0.11 0.10 2 8 80 0.11 0.10 3 9 100 0.10 0.10 4 1 80 0.10 0.10 5 5 100 0.10 0.10 6 50 90 0.10 0.10 7 49 80 0.10 0.10 8 18 100 0.10 0.10 9 16 70.0 0.10 0.11 10 29 90.0 0.10 0.10

(b) Recommended Test Set

Table 9: The Top-10 MRR Content Feature Combination Re-sults

Rank Combination Epoch Mean p@5 Std

1 48 100 0.37 0.35 2 0 100 0.37 0.35 3 16 100 0.37 0.34 4 29 100 0.37 0.34 5 26 100 0.37 0.35 6 19 100 0.36 0.34 7 27 100 0.36 0.34 8 38 100 0.36 0.34 9 9 100 0.36 0.34 10 3 100 0.36 0.34

(a) Test Set

Rank Combination Epoch Mean p@5 Std

1 48 100 0.31 0.32 2 1 90 0.31 0.32 3 51 100 0.30 0.32 4 9 100 0.30 0.32 5 8 80 0.30 0.30 6 16 90 0.30 0.32 7 0 70 0.30 0.33 8 29 100 0.30 0.32 9 20 100 0.29 0.32 10 50 100 0.29 0.31

(b) Recommended Test Set

Referenties

GERELATEERDE DOCUMENTEN

We therefore propose a conceptual framework and the contextualized attention metadata schema that enables the recording and management of rich and detailed sets of data about user

RHCe employs a number of domain experts (cultural heritage experts) whose full- time job is to provide high quality metadata over multimedia documents based on a

Kathleen O’Connor elaborates on the proposal that the language of sexual violation throughout the book of Jeremiah is used to describe the physical and also emotional

Als we er klakkeloos van uitgaan dat gezondheid voor iedereen het belangrijkste is, dan gaan we voorbij aan een andere belangrijke waarde in onze samenleving, namelijk die van

maar dit is in werkelìjkheid het bewerkte portret van Anne-Bobert Turgot zoals geschilderd door Joseph

Overlay The final step in the species mapping procedure is to combine the spatial data sets related to species’ ecological preferences with a fixed rule that determines the final

Kenmerken toevoegen aan leermateriaal is nodig voor de vragen van leraren en leerlingen, maar ook voor systemen die deze vragen adresseren.. Voor die systemen is het gebruik van

If the document-level metadata contains values in the pdfaid namespace, metadata extension schemas will be appended to the document-level metadata packet automatically.