<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:slash="http://purl.org/rss/1.0/modules/slash/">
  <channel>
    <language>en</language>
    <title>Technology + Creativity at the BBC Feed</title>
    <description>Technology, innovation, engineering, design, development.
The home of the BBC's digital services.</description>
    <pubDate>Mon, 04 Oct 2021 15:51:46 +0000</pubDate>
    <generator>Zend_Feed_Writer 2 (http://framework.zend.com)</generator>
    <link>https://www.bbc.co.uk/blogs/internet</link>
    <atom:link rel="self" type="application/rss+xml" href="https://www.bbc.co.uk/blogs/internet/rss"/>
    <item>
      <title>Fighting misinformation: An embedded media provenance specification</title>
      <description><![CDATA[We are looking at technology solutions to the problem of news disinformation. Here is an overview of a draft version of an embedded media provenance specification, released by the Coalition for Content Provenance and Authenticity (C2PA).]]></description>
      <pubDate>Mon, 04 Oct 2021 15:51:46 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/5d371e1b-54be-491f-b8ee-9e354bafb168</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/5d371e1b-54be-491f-b8ee-9e354bafb168</guid>
      <author>Charlie Halford</author>
      <dc:creator>Charlie Halford</dc:creator>
      <content:encoded><![CDATA[<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p09xqbl3.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p09xqbl3.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p09xqbl3.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p09xqbl3.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p09xqbl3.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p09xqbl3.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p09xqbl3.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p09xqbl3.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p09xqbl3.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>For the last few years, the BBC has had a project running in its technology division looking at technology solutions to various problems in the domain of news disinformation. Part of that effort, called <a href="https://www.originproject.info/">Project Origin</a>, is working to make it easier to understand where the news you consume online really comes from so that you can decide how credible it is. You can find some history on this in <a href="https://www.bbc.co.uk/mediacentre/articles/2021/project-origin-one-year-on">Laura&nbsp;Ellis' excellent&nbsp;"Project&nbsp;Origin: one year on"&nbsp;blog</a>.</p>
<p>Part of Project Origin has been working in collaboration with major media and tech companies, most recently, with the&nbsp;<a href="https://c2pa.org">Coalition&nbsp;for Content Provenance and Authenticity&nbsp;(C2PA)</a>,&nbsp;which it helped form. This group recently released <a href="https://c2pa.org/public-draft/">a draft&nbsp;version of an embedded media provenance specification</a>. This spec tackles the problem of missing, trusted provenance information in images / video / audio consumed on the internet. For example, where a video of elections in one country from 10 years ago is being presented as video from recent elections in another.&nbsp;This is&nbsp;an overview of&nbsp;how that specification is intended to work.</p>
<h2 data-usually-unique-id="942583191341384741354902">Embedding</h2>
<p>The C2PA specification works primarily by defining mechanisms for embedding additional data into media assets to indicate their authentic origin. An essential aspect of this data is&nbsp;"assertions"&nbsp;- statements about when and where media was produced. The embedded information is then digitally signed so that a consumer knows who is making the statements.</p>
<p>While the C2PA specification also includes mechanisms for locating this provenance data remotely&nbsp;(e.g.&nbsp;hosted somewhere on the internet), I'll focus on the use case where all data is embedded directly in the asset itself.</p>
<h2 data-usually-unique-id="555473361838427318559989">Data&nbsp;model</h2>
<p>The C2PA specification uses a few different mechanisms for embedding and storing data. Embedding is done with <a href="https://www.iso.org/standard/73604.html">JUMBF</a>,&nbsp;a container format,&nbsp;and structured data storage is done with a combination of <a href="https://www.w3.org/TR/json-ld11/">JSON-LD</a>&nbsp;and&nbsp;<a href="https://www.rfc-editor.org/rfc/rfc8949">CBOR</a> (which&nbsp;is a binary format based on JSON).</p>
<p><strong>Container - The</strong><strong>&nbsp;</strong><strong>"Manifest</strong><strong>&nbsp;Store"</strong></p>
<p>Similar to <a href="https://wwwimages2.adobe.com/content/dam/acom/en/devnet/xmp/pdfs/XMP%20SDK%20Release%20cc-2016-08/XMPSpecificationPart3.pdf">XMP</a>,&nbsp;the C2PA specification defines&nbsp;several&nbsp;embedding points in a selection of media formats to place a&nbsp;"Manifest&nbsp;Store" in JUMBF format, which is the container for the various pieces of provenance data. Once you've identified where and how a manifest store is embedded in your favourite media format, most of the specification is format-agnostic.</p>
<blockquote>
<p><strong>What is JUMBF?</strong></p>
<p>JUMBF&nbsp;(JPEG&nbsp;universal metadata box format) is a binary container format initially designed for adding metadata to JPEG files, and it&rsquo;s now used in other file formats too. It is structurally similar to the <a href="https://en.wikipedia.org/wiki/ISO/IEC_base_media_file_format">ISO&nbsp;Base Media box file format</a>, an extensible container format that is used for many different types of media files. JUMBF&nbsp;"superboxes"&nbsp;are boxes that only contain other boxes. JUMBF&nbsp;"content&nbsp;type" boxes contain actual payload data, the serialisation of which should match the advertised content type of the box. All boxes have labels, which allow boxes to be addressed and understood when parsing. C2PA uses JUMBF in all the media formats it supports to provide the container format for the Manifest, Claims, Assertions, Verifiable Credentials and Signatures.</p>
</blockquote>
<p>Each piece of embedded provenance data is called a&nbsp;&ldquo;Manifest&rdquo;.&nbsp;A manifest contains a part of the provenance data about the current asset, or the assets it was made from. Because an asset might have been created from multiple original sources or have been processed multiple times, we will often need to store several manifests to understand the complete history of the current asset.</p>
<p>Manifests are located in the&nbsp;"Manifest&nbsp;Store", which is a JUMBF superbox. The last manifest in the store is the&nbsp;"ActiveManifest", which is the provenance data&nbsp;about the current asset&nbsp;and it's the logical place for validation to start. The other manifests are the data for the&nbsp;"ingredients"&nbsp;of the active manifest - i.e. the assets that were a part of the creation of the active manifest. This is one of the key features of C2PA: each asset provides a graph of the history of editing and composition actions that went into the active asset, exposing as little or as much as the asset publisher wants.</p>
<p>Each manifest within the store is again its own JUMBF superbox. A manifest then consists of: a&nbsp;"Claim",&nbsp;an&nbsp;"AssertionStore", a&nbsp;"<a href="https://www.w3.org/TR/vc-data-model/">W3C&nbsp;Verifiable Credentials</a>" and a&nbsp;"Signature".&nbsp;Manifests are signed by an actor&nbsp;(the&nbsp;&ldquo;Signer&rdquo;)&nbsp;whose credential identifies them to the user validating or consuming them.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p09x3btj.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p09x3btj.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p09x3btj.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p09x3btj.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p09x3btj.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p09x3btj.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p09x3btj.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p09x3btj.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p09x3btj.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Diagram of a Manifest box, without any VCs</em></p></div>
<div class="component prose">
    <p><strong>Assertions</strong></p>
<p>Assertions are the statements being made by the signer&nbsp;of a manifest. They are the bits of provenance data that consumers of that data are being asked to trust, for example,&nbsp;the date of image capture, the geographical location, or the publisher of a video.<br />In the spec, each assertion has its own data model. Some are published as&nbsp;"Standard&nbsp;Assertions" in the spec, some are adoptions of existing metadata specifications such as&nbsp;<a href="https://www.cipa.jp/std/documents/e/DC-008-2012_E.pdf">EXIF</a>,&nbsp;<a href="https://iptc.org/standards/photo-metadata/">IPTC</a>&nbsp;and&nbsp;<a href="https://schema.org/">schema.org</a>,&nbsp;and it is expected that implementers will extend the spec by defining their own as well.</p>
<blockquote>
<p><strong>Media metadata isn't new</strong></p>
<p>For example, the EXIF standard is nearly universal in digital photographs, used to record location and camera settings. The fundamentally new thing that C2PA does&nbsp;is allow you to cryptographically bind that metadata&nbsp;(with&nbsp;hashes) to a particular media asset and then sign it&nbsp;with the identity credential&nbsp;of the origin of that data, ensuring that the result is tamper-proof and provable.</p>
</blockquote>
<p>Assertions are contained in their own JUMBF Content Type Box in the assertion store superbox&nbsp;and are serialised in the format defined in the spec for that assertion. The C2PA-defined assertions are stored as CBOR, while most adopted assertions from other standards are JSON-LD.</p>
<p>Here's an example of an&nbsp;"Action"&nbsp;assertion&nbsp;(in&nbsp;<a href="https://www.rfc-editor.org/rfc/rfc8949#name-diagnostic-notation">CBOR&nbsp;Diag</a>) which tells you what the signer&nbsp;thinks was done in creating the active asset:</p>
</div>
<div class="component code">
    <pre class="code__pre br-box-subtle"><code class="code__code">{
  &quot;actions&quot;: [
    {
    &quot;action&quot;: &quot;c2pa.filtered&quot;,
    &quot;when&quot;: 0(&quot;2020-02-11T09:00:00Z&quot;),
    &quot;softwareAgent&quot;: &quot;Joe&#039;s Photo Editor&quot;,
    &quot;changed&quot;: &quot;change1,change2&quot;,
    &quot;instanceID&quot;: 37(h&#039;ed610ae51f604002be3dbf0c589a2f1f&#039;)
    }
  ]
}</code></pre>
</div>
<div class="component prose">
    <p>And here's an EXIF one (in JSON-LD) that contains location data:</p>
</div>
<div class="component code">
    <pre class="code__pre br-box-subtle"><code class="code__code">{
  &quot;@context&quot; : {
    &quot;exif&quot;: &quot;http://ns.adobe.com/exif/1.0/&quot;,
  },
  &quot;exif:GPSLatitude&quot;: &quot;39,21.102N&quot;,
  &quot;exif:GPSLongitude&quot;: &quot;74,26.5737W&quot;,
  ...
}</code></pre>
</div>
<div class="component prose">
    <p>The one critical assertion is the binding, something that binds the claim to an asset. In fact, the spec requires one. This ensures that claims are not applied to any asset other than the one they were signed against. This is important in helping to ensure that the consumer can trust that the C2PA data wasn't tampered with between the publisher and the consumer. There are currently two types of "hard bindings" available, a simple hash binding to an area of bytes in a file or a more complex one intended for ISO BMFF-based assets, which can use their box format to reference specific boxes that should be hashed.</p>
<p><strong>Claim</strong></p>
<p>The claim in a manifest exists to pull together the assertions being made, any "redactions" (removals of previous provenance data for privacy reasons), and some extra metadata about the asset, the software that created the claim, and the hashing algorithm used. Assertions are linked by their reference in the assertion store and a hash. The claim itself is another JUMBF box, serialised as a CBOR structure. This is the thing that is signed, and it provides a location to find the signature itself.</p>
<p><strong>Signature</strong></p>
<p>The signature in a manifest is a <a href="https://datatracker.ietf.org/doc/html/rfc8152">COSE</a> CBOR structure that signs the contents of the claim box. COSE is the CBOR version of the JOSE framework of specs, which includes JWT/JWS. The signature is produced using the credentials of the signer. The signer is the primary point of trust in the C2PA Trust Model, and consumers are expected to use the signer's identity to help them make a trust decision on the claim's assertions.</p>
<p>The only currently supported credentials for producing the signature are x.509 certs. The specification provides a profile that certificates are expected to adhere to (including key usages such as &ldquo;id-kp-emailProtection&rdquo;, which is a placeholder). The specification does not include any requirements on how validators &amp; consumers assemble lists of trusted issuers, as it is expected that an ecosystem of issuers will develop around this specification. Instead, it simply requires that validators maintain or reference such a list of trust anchors. Alternatively, they can put together a trusted list of individual entity certificates provided out-of-band of the trust anchor list.</p>
<h2>What now?</h2>
<p>This is an overview and omits both the detail required to produce C2PA manifests and the breadth of some of the other components of the specification (e.g. ingredients, the use of Verifiable Credentials, the concept of assertion metadata, timestamping etc). I'd love to produce a worked example of how to extract and validate a C2PA manifest from an asset; watch out for that in the future. I will highlight an <a href="https://github.com/numbersprotocol/pyc2pa">open-source implementation of C2PA available in Python</a>, and I know of other implementations in the works, too.</p>
<p>At the BBC, we can't wait for this specification to develop and gain adoption. We'd love to see it supported in production and distribution tools, web browsers, and on social media and messaging platforms. We really think it can make a difference to some of the <a href="https://www.bbc.co.uk/news/world-asia-india-47878178">harms done by mis- and disinformation</a>.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Introducing machine-based video recommendations in BBC Sport</title>
      <description><![CDATA[Rob Heap explains how algorithm-based recommendations are saving time in production]]></description>
      <pubDate>Tue, 14 Sep 2021 12:07:03 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/f74ea410-5ec4-4add-9daa-a29d25176ccd</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/f74ea410-5ec4-4add-9daa-a29d25176ccd</guid>
      <author>Robert Heap</author>
      <dc:creator>Robert Heap</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>From this week we are adding a new feature to our short form video pages on the BBC Sport website.</p>
<h4>Related clips</h4>
<p>On every video page in BBC Sport you&rsquo;ll see a related links section. This is usually put together by our editorial colleagues, a routine task which can be time consuming. They have good knowledge about related content, but cannot know about everything, which means that the audience do not see some content that might be relevant.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p09vxrj6.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p09vxrj6.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p09vxrj6.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p09vxrj6.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p09vxrj6.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p09vxrj6.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p09vxrj6.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p09vxrj6.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p09vxrj6.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <h4>Short Form Video &amp; Datalab</h4>
<p>With this in mind we have worked with Datalab (our in-house BBC machine learning specialists) to create an algorithm-based video recommendations engine which we hope will help our audience see more of the content they love whilst reducing the editorial overhead of creating a set of relevant links.</p>
<h4>Algorithm-based recommendations</h4>
<p>The engine works by combining content information about the clip with more information about user journeys from across the BBC. This combination of multiple sources should provide a more relevant list of videos for our audience to watch next. This is the first cross product engine supporting user journeys across News and Sport, which means that you may see news, sport or a combination of both in the module. This is the first version of the short form video recommender, there will be more improvements to come, as we continue to develop it.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p09vxz05.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p09vxz05.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p09vxz05.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p09vxz05.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p09vxz05.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p09vxz05.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p09vxz05.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p09vxz05.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p09vxz05.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <h4>Launching the new recommendations</h4>
<p>The plan is to launch this new functionality on all American Football clips the week commencing 13 September to give the engine a trial with freshly published content and to give us the opportunity to measure its impact. Provided all is well, we will gradually release this feature across all BBC Sport videos. After that we will begin to roll out the same engine for BBC News. Beyond that we will continue to work with editorial colleagues to improve it over the coming months.</p>
<p>If you have any feedback on this new video experience, please leave your comments below.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>How metadata will drive content discovery for the BBC online</title>
      <description><![CDATA[How metadata will be at the heart a new content discovery strategy.]]></description>
      <pubDate>Wed, 15 Apr 2020 13:53:12 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/eacbb071-d471-4d85-ba9d-938c0c800d0b</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/eacbb071-d471-4d85-ba9d-938c0c800d0b</guid>
      <author>Jonathan Murphy and Jeremy Tarling</author>
      <dc:creator>Jonathan Murphy and Jeremy Tarling</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p><em>Jonathan Murphy, editorial lead for metadata and Jeremy Tarling, lead data governance specialist in Digital Publishing, explain what's being done to create a common metadata structure.&nbsp;</em></p>
<p>The BBC&rsquo;s online portfolio has been built up over more than 20 years into a rich and varied collection of websites and services - but with all this content, it&rsquo;s sometimes difficult trying to find some of it, let alone manage it all. As a result we have lots of hidden gems that aren&rsquo;t being surfaced, and that&rsquo;s something that we&rsquo;re trying to fix with a new content discovery strategy.</p>
<p>One of the challenges we&rsquo;re facing up to as we rebuild our digital portfolio is how to make more of our content discoverable and personalised to more of our audiences. That&rsquo;s particularly true of the under 35s age group who now have an array of competing platforms which do a great job in building algorithms to attract their attention.</p>
<p>Underpinning this strategy of discovery and personalisation will be establishing more detailed metadata that describes all of our content using the same terminology and the same tools and data model.</p>
<p>Metadata is the background info that describes the things we make. It can come in all forms, from technical metadata such as which camera was used in a film shoot to promotional metadata used to describe the plot of a programme. For the remit of this project, we&rsquo;re focusing on what we call descriptive content metadata - tags that describe what an asset (e.g. an article, programme or TV/Audio clip) is about or who/what it mentions. That&rsquo;s already used in areas like the BBC News and BBC Sport websites using data architecture called <a href="https://www.bbc.co.uk/blogs/bbcinternet/2012/04/sports_dynamic_semantic.html">Dynamic Semantic Publishing that was created for the London 2012 Olympics,</a> and now drives many thousands of subject-based aggregations, or Topic pages.</p>
<p>There are also <a href="https://www.bbc.co.uk/sounds/categories/mixes">categories of programmes on BBC Sounds</a> and BBC iPlayer, which use a mixture of genres and formats contained in <a href="https://www.bbc.co.uk/blogs/bbcinternet/2009/02/what_is_pips.html">the PIPs database</a> that supports our vast online library of programme information. As a result of these two data silos, and their limitations, it&rsquo;s difficult to offer audiences any pan-BBC experiences or anything that requires an in-depth understanding of the content.</p>
<h4>Common Metadata</h4>
<p>So we need to go further than that, in order to offer material that covers both the breadth of our online content and to suit everyone&rsquo;s tastes and needs. Here in the BBC&rsquo;s Digital Publishing team, we&rsquo;re developing tools that make content description possible at all stages of production across our portffolio, and new vocabularies that allow for richer descriptions of our content.</p>
<p>We&rsquo;ve already worked with the Sounds team, who have used our new tags and curations to create some of their <a href="https://www.bbc.co.uk/sounds/category/mixes">Music Playlists</a> which give you a soundtrack to suit your mood, whether that&rsquo;s &lsquo;chilled out&rsquo; &lsquo;feel good&rsquo;, music to dance to, or music to focus your mind.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p089knwq.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p089knwq.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p089knwq.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p089knwq.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p089knwq.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p089knwq.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p089knwq.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p089knwq.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p089knwq.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>To make this possible we&rsquo;ve developed a concept that every piece of content has a basic set of common metadata associated with it, that it carries around wherever it&rsquo;s surfaced across the BBC&rsquo;s portfolio - whether that&rsquo;s in Sounds, iPlayer or on the BBC News homepage. We&rsquo;re storing this set of basic common metadata in what we call a &lsquo;Passport&rsquo;, and to create and manage this metadata we&rsquo;ve developed a tool called Passport Control.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p089kphl.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p089kphl.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p089kphl.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p089kphl.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p089kphl.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p089kphl.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p089kphl.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p089kphl.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p089kphl.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>Using the simple graph model first developed for the BBC&rsquo;s 2012 Olympic&rsquo;s coverage, we create subject-predicate-object triples to describe the nature of the relationship between an asset (the subject) and a tag (the object).</p>
<p>In the above example this asset has been described as being &ldquo;about&rdquo; some things - this kind of subject based tagging is well established at the BBC, especially in our journalism output. But we have added three new predicates: &ldquo;editorial tone&rdquo;, &ldquo;intended audience&rdquo;, and &ldquo;genre&rdquo;.</p>
<p>Each predicate can be used with an associated controlled vocabulary of terms. In some cases these controlled vocabularies are taxonomic hierarchies (like BBC genres) while in others they are simple lists of terms that we have developed to describe our output in ways that make sense to us and our audience.</p>
<p>These new types of metadata can be used to make much richer collections of content, either as manual editorial curations or algorithmically generated recommendations.</p>
<p>Our colleague Anna McGovern explains further some of those challenges we face at the BBC <a href="https://www.bbc.co.uk/blogs/internet/entries/887fd87e-1da7-45f3-9dc7-ce5956b790d2">here</a> in building our curations and recommendations, building on our public service values. With the amount and variety of material that we produce, from news articles to music mixes, live events to boxsets, we think we&rsquo;re in a good position to provide content for all kinds of different tastes.</p>
<p>We&rsquo;ll update you more about metadata developments, curations and recommendations as these features begin to roll out on BBC Online over the coming months.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Scaling responsible machine learning at the BBC</title>
      <description><![CDATA[How the BBC's public service principles are being applied to machine learning.]]></description>
      <pubDate>Fri, 04 Oct 2019 09:32:44 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/4a31d36d-fd0c-4401-b464-d249376aafd1</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/4a31d36d-fd0c-4401-b464-d249376aafd1</guid>
      <author>Gabriel Straub</author>
      <dc:creator>Gabriel Straub</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>Machine learning is a set of techniques where computers can &lsquo;solve&rsquo; problems without being explicitly programmed with all the steps to solve the problem, within the parameters set and controlled by data scientists working in partnership with editorial colleagues.</p>
<p>The BBC currently uses machine learning in a range of ways &ndash; for example to provide users with personalised content recommendations, to help it understand what is in its vast archive, and to help transcribe the many hours of content we produce. And in the future, we expect that machine learning will become an ever more important tool to help the BBC create great audience experiences.</p>
<p>The BBC was founded in 1922 in order to inform, educate and entertain the public. And we take that purpose very seriously. We are governed by our <a href="https://www.bbc.com/aboutthebbc/governance/charter">Royal Charter</a>&nbsp;and public service is at the heart of everything we do. This means that we act on behalf of our audience by giving them agency and that our organisation exists in order to serve individuals and society as a whole rather than a small set of stakeholders.</p>
<p>With Machine Learning becoming a more prevalent aspect of everyday life, our commitment to audience agency is reflected in this area as well. And so in 2017, we submitted <a href="http://data.parliament.uk/writtenevidence/committeeevidence.svc/evidencedocument/artificial-intelligence-committee/artificial-intelligence/written/70493.pdf%20">a written commitment to the House of Lords Select Committee on Artificial Intelligence</a>&nbsp;in which we promised to be leading the way in terms of responsible use of all AI technologies, including machine learning.</p>
<p>But what does this mean in practice?</p>
<p>For the last couple of months, we have been bringing together colleagues from editorial, operational privacy, policy, research and development, legal and data science teams in order to discuss what guidance and governance is necessary to ensure our machine learning work is in line with that commitment.</p>
<p>Together, we agreed that the BBC&rsquo;s machine learning engines will support public service outcomes (i.e. to inform, educate and entertain) and empower our audiences.</p>
<p>This statement then led to a set of <strong>BBC Machine Learning Principles</strong>:</p>
<h4>The BBC&rsquo;s Values</h4>
<p>1. The BBC&rsquo;s ML engines will reflect the values of our organisation; upholding trust, putting audiences at the heart of everything we do, celebrating diversity, delivering quality and value for money and boosting creativity.</p>
<h4>Our Audiences</h4>
<p>2. Our audiences create the data which fuels some of the BBC&rsquo;s ML engines, alongside BBC data. We hold audience-created data on their behalf, and use it to improve their experiences with the BBC.</p>
<p>3. Audiences have a right to know what we are doing with their data. We will explain, in plain English, what data we collect and how this is being used, for example in personalisation and recommendations.</p>
<h4>Responsible Development of Technology</h4>
<p>4. The BBC takes full responsibility for the functioning of our ML engines (in house and third party). Through regular documentation, monitoring and review, we will ensure that data is handled securely. And that our algorithms serve our audiences equally and fairly, so that the full breadth of the BBC is available to everyone.</p>
<p>5. Where ML engines surface content, outcomes are compliant with the BBC&rsquo;s editorial values (and where relevant as set out in our editorial guidelines). We will also seek to broaden, rather than narrow, our audience&rsquo;s horizons.</p>
<p>6. ML is an evolving set of technologies, where the BBC continues to innovate and experiment. Algorithms form only part of the content discovery process for our audiences, and sit alongside (human) editorial curation.</p>
<p>These principles are supported by a checklist that gives practitioners concrete questions to ask themselves throughout a machine learning project. These questions are not formulated as a governance framework that needs to be ticked off, but instead aim to help teams building machine learning engines to really think about the consequences of their work. Teams can reflect on the purpose of their algorithms; the sources of their data; our editorial values; how they trained and tested the model; how the models will be monitored throughout their lifecycle and their approaches to security and privacy and other legals questions.</p>
<p>While we expect our six principles to remain pretty consistent, the checklist will have to evolve as the BBC develops its machine learning capabilities over time.</p>
<p>The <a href="https://findouthow.datalab.rocks/">Datalab team</a>&nbsp;is currently testing this approach as they build the BBC&rsquo;s first in-house recommender systems, which will offer a more personalised experience for BBC Sport and BBC Sounds. We also hope to improve the recommendations for other products and content areas in the future. We know that this framework will only be impactful if it is easy to use and can fit into the workflows of the teams building machine learning products.</p>
<p>The BBC believes there are huge benefits to being transparent about how we&rsquo;re using Machine Learning technologies. We want to communicate to our audiences how we&rsquo;re using their data and why. We want to demystify machine learning. And we want to lead the way on a responsible approach. These factors are not only essential in building quality ML systems, but also in retaining the trust of our audiences.</p>
<p>This is only the beginning. As a public service, we are ultimately accountable to the public and so are keen to hear what you think of the above.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Navigating the data ecosystem technology landscape</title>
      <description><![CDATA[The challenges of creating an open and transparent data ecosystem for the BBC.]]></description>
      <pubDate>Tue, 03 Sep 2019 12:46:36 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/67fee994-3d20-45d5-be2a-acfc47d572f1</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/67fee994-3d20-45d5-be2a-acfc47d572f1</guid>
      <author>Hannes Ricklefs, Max Leonard</author>
      <dc:creator>Hannes Ricklefs, Max Leonard</dc:creator>
      <content:encoded><![CDATA[<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p07mb9b5.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p07mb9b5.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p07mb9b5.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p07mb9b5.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p07mb9b5.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p07mb9b5.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p07mb9b5.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p07mb9b5.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p07mb9b5.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Credit: Jasmine Cox</em></p></div>
<div class="component prose">
    <p>Want to message your Facebook friends on Twitter? Move your purchased music from iTunes to Amazon? Get Netflix recommendations based on your iPlayer history? Well, currently you can&rsquo;t.</p>
<p>Many organisations are built on data, but the vast majority of the leading players in this market are structured as vertically integrated walled gardens, with few (if any) meaningful interfaces to any outside services. There are a great number of reasons for this, but regardless of whether they are intentional or technological happenstance (or a mixture of both), there is a rapidly growing movement of GDPR supercharged technologists who are putting forward <a href="https://www.bbc.co.uk/news/technology-45706429">decentralised and open alternatives</a> to the <a href="https://www.intricity.com/data-science/what-is-a-data-moat/">data-moated</a> household names of today. For the BBC in particular, these new ways of approaching data are well aligned with our public service ethos and commitment to treating data in the most ethical way possible.</p>
<p>Refining how the BBC uses data, both personal and public, is critical if we are to create <a href="https://www.bbc.co.uk/mediacentre/speeches/2017/tony-hall-annual-plan#heading-a-personalised-uniquely-tailored-bbc">a truly personalised BBC</a> in the near term and essential if we want to remain relevant in the coming decades. Our Chief Technology and Product Officer Matthew Postgate <a href="https://www.bbc.co.uk/blogs/internet/entries/78948980-e1e6-48fe-918a-c9bb5f2a0719">recently spoke about the BBC&rsquo;s role within data-led services</a>, in which he outlined some of the work we have been doing in this respect to ensure the BBC and other public service organisations are not absent from new and emerging data economies.</p>
<p>Alongside focused technical research projects like the <a href="https://www.bbc.co.uk/rd/blog/2019-06-bbc-box-personal-data-privacy">BBC Box</a>, we have been mapping the emerging players, technologies and data ecosystems to further inform the BBC&rsquo;s potential role in this emerging landscape. Our view is that such an ecosystem is made up of the following core capabilities: Identity, data management (storage, access, and processing), data semantics and the developer experience, which are currently handled wholesale in traditional vertical services. A first step for us is hence to ascertain which of these core capabilities can realistically be deployed in a federated, decentralised future, and which implementations currently exist to practically facilitate this.</p>
<p><strong>Identity</strong>, a crucial component of the data ecosystem, proves who users say they are providing a true digital identity. Furthermore we expect standard account features such as authentication and sharing options via unique access token that could enable users to get insights or to share data to be part of any offering. We found that identity, in the context of proving a user&rsquo;s identity, was not provided by any of the solutions we investigated. Standard account features were present, ranging from platform specific implementations, to decentralised identifier approaches via WebID, and blockchain based distributed ledger approaches. As we strongly believe it is important to prove a user is who they say they are, at this point we would look to integrate solutions that specialise in this domain.</p>
<p><strong>Data management</strong> can be further broken down into 3 areas:</p>
<ol>
<li><strong>Data usage and access</strong>, involves providing integration of data sources with an associated permission and authorisation model. Users should have complete governance of their data and usage by data services. Strong data security controls and progressive disclosure of data are key here. Given our investigation is based around personal data stores (PDS) and time series sensor/IoT device data platforms to capture personal, public and open data, providing access and controls around sharing of data was a fundamental capability of all offerings. All of them provided significant granularity and transparency to the users about what data is being stored, its source and usage by external services.</li>
<li><strong>Data storage</strong> must provide high protection guarantees of users&rsquo; data, encrypted in transit and at rest, giving users complete control and transparency of data lifecycle management. Again, this is a fundamental requirement, such that storage is either a core offering of any platform or outsourced to external services that store data in strongly encrypted formats.</li>
<li><strong>Data processing</strong> mechanisms to allow users to bring &ldquo;algorithms&rdquo; to their data, combined with a strong contract based exchange of data. Users are in control and understand what insights algorithms and services derive from their data. These might include aspects such as the creation of reports, creation and execution of machine learning models, other capabilities that reinforce the user&rsquo;s control over how their personal data is used for generated insights. Through contract and authorisation based approaches users have complete audit trails of any processing performed which provides transparency of how data is utilised by services, whilst continuously being able to detect suspicious or unauthorised data access. Our investigations found that processing of data is either through providing SDKs that heavily specify the workflow for data processing, or no provisioning at all, leaving it to developers to create their own solution.</li>
</ol>
<p><strong>Data model and semantics&nbsp;</strong>refers to mechanisms that describe (schemas, ontologies) and maintain the data domains inside of the ecosystem, which is essential to provide extensibility and interoperability. Our investigations found this being approached in a wide spectrum from:</p>
<ol>
<li>no provision requiring developers to come to conclusions about the best way to proceed</li>
<li>using open standards such as schema.org and modeling data around linked data and RDF</li>
<li>completely proprietary definitions around schemas within the system.</li>
</ol>
<p>Finally the <strong>developer experience</strong> is key. It requires a set of software development tools to enable engineers to develop features and experiences as well as being able to implement unique value propositions required by services. This is the strongest and most consistent area across all our findings.</p>
<p>In summary our investigations have shown that there is no one solution that provides all of our identified and required capabilities. Crucially the majority of the explored end user solutions are still commercially orientated, such that they either make money from subscribers or through associated services.</p>
<p>So with the number of start-ups, software projects and standards that meet these capabilities snowballing, where might the BBC fit into this increasingly crowded new world?</p>
<p>We believe that the BBC has a role to play in all of these capabilities and that it would enhance our existing public service offering: to inform, educate and entertain. A healthy ecosystem requires multiple tenants and solutions providers, all adhering to core values such as transparency, interoperability and extensibility. Only then will users be able to freely and independently move or share their data between providers which would enable purposeful collaboration and fair competition toward delivering value to audiences, society and industry.</p>
<p>The BBC was incorporated at the dawn of the radio era to counteract the unbridled free-for-all that often comes with any disruptive technology, and <a href="https://www.bbc.co.uk/rd/about/our-purpose">its remit to shape standards and practices </a>for the good of the UK and its population stands today as <a href="http://downloads.bbc.co.uk/historyofthebbc/1920s.pdf">it did in 1927</a>. With a scale, reach and purpose that is unique to the BBC, it is strongly congruent with our public service duty to help drive policy, standards and access rights to ensure that the riches on offer in these new ecosystems are not coopted solely for the downward pursuit of profit, and remain accessible for the benefit of all.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Machine learning and editorial collaboration within the BBC</title>
      <description><![CDATA[Datalab team members explain how the best machine learning results will come from a multi-disciplined approach.]]></description>
      <pubDate>Thu, 29 Aug 2019 13:55:00 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/a38207dd-e4ed-40fa-8bdf-aebe1dc74c28</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/a38207dd-e4ed-40fa-8bdf-aebe1dc74c28</guid>
      <author>Anna McGovern, Ewan Nicolson, Svetlana Videnova</author>
      <dc:creator>Anna McGovern, Ewan Nicolson, Svetlana Videnova</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>The BBC is nearly 100 years old. Inevitably, as an organisation we are having to adapt to meet some of the technological requirements of the future, such as incorporating Machine Learning (ML) technologies. ML recommendations, for example, is a standard way for audiences to discover content and the BBC is committed to make this discovery more personal. Developing these services has brought an interesting opportunity for collaboration between the ML and Editorial teams within Datalab, the BBC team focused on building recommendation engines.</p>
<p>About a year ago we started the experiment of the <a href="https://www.bbc.co.uk/blogs/internet/entries/82cd8d1e-2f23-4eff-8f34-0ef38ca8854c">BBC+ app</a>. This was the first time the BBC provided the audience with a fully automated ML service. With this wealth of knowledge and with more data science initiatives taking shape, we want to use all the available expertise the BBC can provide.</p>
<p>Our aim is to create responsible recommendation engines, true to the BBC values and using all available expertise the BBC can provide. In industry, it is commonplace for data science teams to make use of specialist knowledge to inform how models are developed. For example, data scientists working for a travel site would use experts with knowledge about everything from business flights to how and when families go on holiday. Datalab consulted editorial teams and representatives who specialised in curation as it began to develop recommendations for content discovery.</p>
<p>Datalab&rsquo;s editorial lead, Anna McGovern, helps us with advice on editorial judgement and content curation expertise within the BBC. Ewan Nicolson is lead data scientist and represents the technological aspect of Datalab&rsquo;s work here. Svetlana Videnova, Business Analyst, poses some of the common teamwork problems within the public media industry and technological challenges we face today. We will focus on a given challenge about the curation of the content and leave its creation phase for another post. Both Anna and Ewan will provide their way of tackling that work in their own fields. The last column of the table below demonstrates an example of how the collaboration works in our team.</p>
<p>As you&rsquo;ll see, the two fields of editorial and data science compliment each other. Working across discipline gives better results for the audience, and helps us learn from each other. It means that machine learning is actually solving the correct problems because we&rsquo;re making use of the rich expertise from editorial. It also means that editorial are able to take advantages of techniques like machine learning to multiply their efforts and deliver more value to the audience.</p>
</div>
<div class="component prose">
    <table width="554" border="1" cellpadding="0">
<tbody>
<tr>
<td valign="top" width="98">
<p><strong>Challenge</strong></p>
</td>
<td valign="top" width="126">
<p><strong>Machine Learning solution</strong></p>
</td>
<td valign="top" width="147">
<p><strong>Editorial solution</strong></p>
</td>
<td valign="top" width="175">
<p><strong>When we collaborate</strong></p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How&nbsp;do&nbsp;we ensure curation is a good experience for users?</p>
</td>
<td valign="top" width="126">
<p>We consider many different measures of success: accuracy, diversity, recency, impartiality, editorial priority.</p>
</td>
<td valign="top" width="147">
<p>Traditionally on an editorial team, a journalist would research a story, discuss how it might be covered and compose the story itself to make it compelling.&nbsp;</p>
</td>
<td valign="top" width="175">
<p>The data scientists get a rich understanding from editorial of the different trade-offs between these measures of success. Deep domain knowledge.</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How does recency impact curation of content?</p>
</td>
<td valign="top" width="126">
<p>We include publication date as a feature in our models. We sometimes try and optimise for recency, showing people more current content in some situations.</p>
</td>
<td valign="top" width="147">
<p>One of the challenges is that once that work is done it is fairly hard to bring the editorial creation back to life, especially for evergreen content. This is one of many examples that ML recommendations could help&nbsp;with, by surfacing this content in the most relevant time according to the user&rsquo;s experience or history.&nbsp;</p>
</td>
<td valign="top" width="175">
<p>By working together we&rsquo;re able to identify how to make decisions about which pieces of content are evergreen and suitable for recommendation, and which pieces have a limited shelf-life and shouldn&rsquo;t be presented to users beyond a certain point.</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How does the BBC ensure impartiality?&nbsp;</p>
</td>
<td valign="top" width="126">
<p>We use&nbsp;<a href="https://5harad.com/papers/fair-ml.pdf">measures of statistical fairness</a>&nbsp;to understand if our model is giving unbiased results.</p>
<p>&nbsp;</p>
<p>Good practice in machine learning make sure that we&rsquo;re using unbiased training data.</p>
<div align="center"><hr align="center" size="1" width="69" /></div>
</td>
<td valign="top" width="147">
<p>Editors,&nbsp;journalists and content creators make a concerted effort&nbsp;to ensure that a range of views and perspectives are shown within a piece of content or across several pieces of content(within&nbsp;a series for example)</p>
</td>
<td valign="top" width="175">
<p>We combine our good practices with domain knowledge from editorial. We use techniques like&nbsp;human-in-the-loop machine learning, or semi-supervised learning&nbsp;to make editorial&rsquo;s lives easier, and apply their knowledge at massive scale.</p>
<p>&nbsp;</p>
<p>ML helps editorial&nbsp;identifying those pieces of content that show a breadth of views.&nbsp;</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How we ensure&nbsp;variety&nbsp;within content serving?</p>
</td>
<td valign="top" width="126">
<p>We construct mathematical measures for&nbsp;<a href="http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.221.6277&amp;rep=rep1&amp;type=pdf">novelty and diversity.</a>&nbsp;We include these in our machine learning optimisations.</p>
<p>&nbsp;</p>
</td>
<td valign="top" width="147">
<p>Editorial staff responsible for curation&nbsp;ensure&nbsp;a breadth and depth of content on indexes, within collections etc</p>
</td>
<td valign="top" width="175">
<p>We learn about the differences between our different pieces of content. Working together we&rsquo;re able to determine if our recommendations offer an interesting, relevant, and useful journey for the user.&nbsp;</p>
<p>&nbsp;</p>
<p>The BBC&rsquo;s audio networks feature different output and tone of voice. ie.&nbsp;Radio 4 has a very different &lsquo;flavour&rsquo;&nbsp;to 6Music. Consequently network can be used to ensure variety in results.</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How do we avoid&nbsp;legal issues?&nbsp;</p>
</td>
<td valign="top" width="126">
<p>We are given a checklist, and we check the items off. We get told that there are things&nbsp;&ldquo;we can&rsquo;t do for opaque legal reasons&rdquo; but never really understand why, and limit the functionality of our solution.</p>
<p>&nbsp;</p>
</td>
<td valign="top" width="147">
<p>Editors, journalists and content creators have to attend a mandatory course relating to media law, so that they have full knowledge about issues such as contempt of court, defamation and privacy. An editor will sign off content to ensure that content is compliant with legal requirements.&nbsp;</p>
</td>
<td valign="top" width="175">
<p>By talking to legal advisers we can build business rules to minimise the risk of legal infractions.&nbsp;</p>
<p>&nbsp;</p>
<p>Close collaboration with editorial means we gain a deep understanding of the potential problems ahead at an early stage. We build with awareness of these concerns, and with that awareness build a solution that is high quality from both a technical and editorial point of view.</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How we&nbsp;handle editorial quality?</p>
</td>
<td valign="top" width="126">
<p>We build and refine a model using data science good practices, and then turn it over to our editorial colleagues. They then decide if the results are good or not.</p>
</td>
<td valign="top" width="147">
<p>When editors curate they can choose content that is relevant, interesting and of good quality.&nbsp;</p>
<p>&nbsp;</p>
<p>&nbsp;</p>
<p>Recommendations present a specific editorial challenge,&nbsp;in that recommenders can surface content that is not the best of our output.&nbsp;</p>
</td>
<td valign="top" width="175">
<p>In&nbsp;BBC+&nbsp;we prioritised content that we knew would suit the environment in which it appeared: standalone, short-form videos, appearing in a feed, from digital first areas such as Radio 1, The Social, BBC Ideas etc</p>
<p>&nbsp;</p>
<p>Including editorial throughout the process means that they teach us about what is important in the results, so that data science understand the real problems that we&rsquo;re trying to solve.</p>
<p>&nbsp;</p>
<p>We fail quickly, and learn quickly, getting to a better quality result.</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How we learn from our audiences? Accuracy/user generated content?</p>
</td>
<td valign="top" width="126">
<p>Measure user activity with the products, and construct measurements of engagement.</p>
<p>&nbsp;</p>
<p>Building implicit and explicit feedback loops. An explicit feedback loop is having a&ldquo;like&rdquo;&nbsp;button, an implicit feedback loop is determining a way to measure when something has gone wrong, like bounce rate or user churn.</p>
<p>&nbsp;</p>
</td>
<td valign="top" width="147">
<p>We monitor feedback and analyse stats to build a picture about how our audiences engage with our content.&nbsp;</p>
</td>
<td valign="top" width="175">
<p>We work with editorial to understand the insights we get from data. They help rationalise the behaviours that we see in the data. They also teach us things that we should look for in the data.</p>
</td>
</tr>
<tr>
<td valign="top" width="98">
<p>How we test recommendations</p>
</td>
<td valign="top" width="126">
<p>A mixture of offline evaluation metrics&nbsp;(e.g.testing against a known test set of data), and online evaluation metrics&nbsp;(e.g.A/B testing)</p>
<p>&nbsp;</p>
</td>
<td valign="top" width="147">
<p>Traditionally: We monitor feedback and analyse stats to build a picture about how our audiences engage with our content.&nbsp;</p>
</td>
<td valign="top" width="175">
<p>The editorial lead works with data scientists on the composition of the recommender. The results are then reviewed by the editorial lead and to obtain a variety of opinions the results are reviewed by more editorial colleagues.&nbsp;</p>
<p>&nbsp;</p>
<p>More on quantitative testing&nbsp;<a href="https://www.bbc.co.uk/blogs/internet/entries/3e4342d4-6f81-47c0-8ba2-8dc7b419eb72">here</a><strong>&nbsp;.</strong></p>
<p>&nbsp;</p>
<p>The rich editorial feedback lets us understand where our model could be better and make improvements.</p>
</td>
</tr>
</tbody>
</table>
</div>
<div class="component prose">
    <p>We&rsquo;re big believers in cross-disciplinary collaboration. As we&rsquo;ve touched on in this article the BBC has a lot of uniquely complex problems to solve in this space. This collaboration is essential if we&rsquo;re going to continue to deliver value to the BBC&rsquo;s audience using data.</p>
<p>If you are curious about this collaboration and would like to know more in depth about how we work, leave us a message and we will be happy to get back to you.</p>
<p>Also, we are hiring <a href="https://findouthow.datalab.rocks/">https://findouthow.datalab.rocks/</a>.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Developing personalised recommender systems at the BBC</title>
      <description><![CDATA[Some of the approaches that are being tested to improve recommendations.]]></description>
      <pubDate>Thu, 22 Aug 2019 14:20:43 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/3e4342d4-6f81-47c0-8ba2-8dc7b419eb72</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/3e4342d4-6f81-47c0-8ba2-8dc7b419eb72</guid>
      <author>Jana Eggink</author>
      <dc:creator>Jana Eggink</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>The BBC is on a journey to become more personalised, and recommendations are an important part of that goal. To date, recommendations in the BBC have been provided primarily by external providers. We feel that offering &mdash; and understanding &mdash; good recommendations is a crucial area for us in reaching our target audience of young listeners and so we have started exploring this area in-house. The <a href="https://findouthow.datalab.rocks">Datalab team</a> is a relatively new team specialising in machine learning, and looking after recommender systems in the BBC. We work with product groups to develop new ways to personalise their offerings, and also collaborate with <a href="https://www.bbc.co.uk/rd">BBC R&amp;D</a>.</p>
<p>We want to be able to explain the composition of our recommendations and so we need to understand how they are generated. Our recommendations should reflect the breadth and diversity of our content and meet our editorial guidelines, as well as informing, educating and entertaining! All these were good reasons for us to build the capability to constantly create challengers to the existing recommendation models.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p07l9x41.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p07l9x41.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p07l9x41.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p07l9x41.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p07l9x41.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p07l9x41.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p07l9x41.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p07l9x41.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p07l9x41.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Current recommendations on Sounds website</em></p></div>
<div class="component prose">
    <p>Datalab was assigned this brilliant and fun challenge and began collaborating with the Sounds team, using a multidisciplinary group made up of data scientists, engineers, editorial specialists and product managers.</p>
<p>The team had some prior experience building personalised recommendations for our video clip <a href="https://www.bbc.co.uk/blogs/internet/entries/82cd8d1e-2f23-4eff-8f34-0ef38ca8854c">app BBC+</a>. For BBC+, the recommender was purely content based, using existing metadata information such as genres (e.g. Drama/Medical) or brands (e.g. Glastonbury Festival). This would probably have been a good approach if our content had been labelled for the express purpose of personalisation. However, the BBC&rsquo;s production workflows were designed to meet the needs of broadcast systems, and we didn&rsquo;t always have all the labels we would have wanted for recommendations.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p07l9xqd.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p07l9xqd.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p07l9xqd.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p07l9xqd.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p07l9xqd.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p07l9xqd.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p07l9xqd.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p07l9xqd.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p07l9xqd.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>BBC+ app</em></p></div>
<div class="component prose">
    <p><a href="https://www.csie.ntu.edu.tw/~b97053/paper/Rendle2010FM.pdf">Factorisation Machines</a> come with the enticing promise of combining content-based recommendations with collaborative filtering.</p>
<p>Using a standard content-based approach, if a user had listened to podcasts from the genre &lsquo;Health &amp; Wellbeing&rsquo; the system would recommend a new episode from Radio 1&rsquo;s Life Hacks but it could also recommend Radio 4&rsquo;s Inside Health, which has a very different tone of voice. By contrast, collaborative filtering matches programmes based on what similar users have enjoyed &mdash; so if they listen to Radio 1&rsquo;s Life Hacks, they might be recommended Radio 1 comedy. This model relies on &lsquo;adjacent&rsquo; content similar to the recommendations found in shopping websites where &lsquo;customers who bought this also bought that&rsquo;. This approach often leads to better recommendations for established content, but is less effective for fresh content that hasn&rsquo;t been consumed by significant numbers of people. Since the BBC continuously produces new content throughout the day this recommendation strategy by itself would be limiting.</p>
<p>Factorisation machines are a smart way to combine both. They have been around a few years, and open source toolboxes exist to support them. Our team programs primarily in Python, so we wanted a toolbox that integrates with that. Obviously, we also wanted it to be fast, give superior results and be easy to use (more on that later&hellip;).</p>
<p>We stored user-item interactions (i.e. the programmes a specific user has listened to) in a BigQuery table. The programme items with the corresponding genre and brand metadata were in a different table, and both needed to be assembled in the correct format for the factorisation machines. Our first choice of toolbox was xlearn. The code seemed relatively mature, running a first test example was easy, and the toolbox offers a variety of different flavours in terms of learning algorithm. But it was hard to get the data into the correct format and, even now that we have a version up and running, we&rsquo;re still not sure we got everything right &mdash; mainly because the initial results are nowhere near as good as we had wanted (and expected) them to be!</p>
<p>The quality of recommendations can be subjective and we needed a way to test them before making them live anywhere on the BBC&rsquo;s websites or apps. Predicting past behaviour is one way of doing this, but also comes with all sorts of problems: users only click on what they see, a piece of content might be brilliant, but if it does not appear in the results, the user will not see it and cannot click on it. Recommending the most popular items generally gives good numbers (as by definition these items get the most clicks), but rarely leads to recommendations of fresh content. In practical terms, it&rsquo;s also a lot of work to set up if your data is stored in ways that were not devised with the ease of access for data scientists in mind&hellip;</p>
<p>So we decided to test the results using qualitative evaluation, asking about 20 editorial and other non-technical people to judge our new recommendations against those from an existing provider. We didn&rsquo;t tell them which set came from which recommender! We used the individual history of the internal test participants to generate the recommendations by both providers and asked for their preference and general feedback.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p07lbl35.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p07lbl35.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p07lbl35.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p07lbl35.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p07lbl35.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p07lbl35.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p07lbl35.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p07lbl35.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p07lbl35.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Qualitative experiment</em></p></div>
<div class="component prose">
    <p>Most of our test users preferred the recommendations we currently have live to our first set of test recommendations and we weren&rsquo;t keen on them either, so we knew we had more work to do.</p>
<p>With the overall infrastructure set-up, it was quite easy to swap out the toolbox we&rsquo;ve used for the factorisation machines. We had previously looked at <a href="https://lyst.github.io/lightfm/docs/home.html">lightFM,</a> and it had a much simpler data format, so we decided to give it a go. We were able to compute new recommendations and run another qualitative experiment in less than two weeks. Our recommendations looked much better, and our test users agreed! However, these are still first results. We don&rsquo;t feel we&rsquo;ve fully solved the problem of recommending popular items versus programmes that are strongly tailored towards a specific user&rsquo;s interests, and are looking into ways to improve this.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p07lbmwq.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p07lbmwq.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p07lbmwq.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p07lbmwq.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p07lbmwq.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p07lbmwq.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p07lbmwq.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p07lbmwq.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p07lbmwq.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>We are happy with the results so far, but there is still a lot of work to do to bring the recommender into production. The infrastructure needs decidedly more work to make it robust and able to scale, and we&rsquo;d like to do more testing. Having a variety of offline metrics should help us to optimise parameters, and test new algorithms without having to go back to our testing panels every few days. We&rsquo;re also still looking at a simple content-based recommender to have another baseline, so more results hopefully soon.</p>
<p>We also still have some more fundamental questions that we hope our practical work will help us to answer. For example, can we use the same approach for recommending entertainment as for news, or do we need specialised systems for each domain? And what if we change the medium and move from audio and video to text, or new interfaces like voice controlled devices? Even if the overall editorial guidelines do not change, we might need different technical approaches to be able to achieve them. But we also want to avoid starting from scratch for every new recommender we build, and we&rsquo;re still trying to figure out how best to do that. In summary, there is lots to do, but it&rsquo;s exciting and we&rsquo;re enjoying the challenge!</p>
<p>Want to work with us? <a href="https://findouthow.datalab.rocks/">https://findouthow.datalab.rocks/</a></p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Looking at the BBC's role in data-led services</title>
      <description><![CDATA[What the BBC is doing with data services and how it's ensuring this maintains trust with audiences.]]></description>
      <pubDate>Wed, 19 Jun 2019 08:20:47 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/78948980-e1e6-48fe-918a-c9bb5f2a0719</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/78948980-e1e6-48fe-918a-c9bb5f2a0719</guid>
      <author>Matthew Postgate</author>
      <dc:creator>Matthew Postgate</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>It&rsquo;s been a busy time in my team over the last few months &ndash; with updates to BBC Sounds and iPlayer, 5G trials in Orkney (and in London), UHD trials for the FA Cup, Doctor Who launching in Virtual Reality and those teams behind the scenes that keep our broadcast and online services going day-after-day.</p>
<p>But one area that keeps on coming up when I&rsquo;m out and about speaking at conferences, or at meetings with our partners - is the question of data &ndash; how we use it, how we share it and its potential to help us understand the world around us.</p>
<p>Not a week goes by without stories about data. There are negative stories, about data being used to target you with specific messages or sell you more, or leaks of personal data to third-parties. But there are also positive stories, like <a href="https://www.bbc.co.uk/news/uk-scotland-48556413">using big data to help reduce carbon emissions </a>or <a href="https://www.bbc.co.uk/news/technology-48072164">helping the justice system work better</a>.</p>
<p>This has made me think about the BBC&rsquo;s role in this new &lsquo;data economy&rsquo; &ndash; and what that should be.</p>
<h4>How we use your data</h4>
<p>At the BBC, we use data to make what we provide you, our viewers, listeners or readers, even better. It helps us tailor our products and services to be more about you &ndash; recommending programmes or content we think you might like, or alert you to the fact your favourite sport team has just scored (or lost a match). We also use it to ensure we&rsquo;re making something for all audiences &ndash; and helps find gaps when we commission programmes and services.</p>
<p>But is there more that we could be doing to ensure data is used for good &ndash; that the data you give organisations is not just used for commercial gain but is used in a way that helps you and potentially your wider community? We think, potentially, yes.</p>
<p>That&rsquo;s why we&rsquo;ve started to work with teams here at the BBC and other partners on specific projects to help us identify what public service value we can bring to these new markets driven by data.</p>
<p>To be clear &ndash; we&rsquo;re experimenting at this stage, and we will learn what works, what people might like &ndash; and what areas we think the BBC can help with, as we go along. We&rsquo;re particularly interested in learning about how organisations can share data to get new insights and how people can safely move their data around. And, we know that when it comes to data, people are rightly concerned about privacy, safety and security. That&rsquo;s why these trials will start small and controlled, so participants will have signed-up clear in knowing what, why and how their data is being used.</p>
<h4>So what have we been up to?</h4>
<p>Late last year, the DCMS <a href="https://www.gov.uk/government/publications/research-on-data-portability">published a report</a> which looked at the potential of personal data portability to stimulate innovation and competition in the UK. It found that the ability to safely and securely move personal data around could unlock huge economic and societal gains, but that there are big practical issues (both in the way organisations share data and how consumers use it) to resolve first.</p>
<p>Following this, (and with DCMS, ICO and CDEI as observers), we&rsquo;re involved in two controlled trials of data sharing by 25 individuals. These trials tests how it could be practicality possible to put a person in control of the data they share about themselves with other companies and what concerns this brings up.</p>
<p>The first trial is cross-sector, with the participants signing up to share data from a range of commercial companies &ndash; as well as the BBC. You can find out more about that <a href="https://www.ctrl-shift.co.uk/news/2019/06/17/release-of-data-mobility-infrastructure-sandbox-report">here</a>.&nbsp;</p>
<p>The second looks at bringing together data from media providers into a BBC data store (or what we&rsquo;re calling internally a BBC Box) to improve people&rsquo;s experiences when watching or listening to programmes. Bill has blogged about this <a href="https://www.bbc.co.uk/rd/blog/2019-06-bbc-box-personal-data-privacy">here</a>.</p>
<h4>What&rsquo;s next?</h4>
<p>Over the coming months, we&rsquo;ll continue looking at this area &ndash; with more experiments and closed trials.</p>
<p>We&rsquo;ll be sharing more about what we learn &ndash; and look at what value the BBC can bring you &ndash; ensuring this market develops in a way that maximises the huge potential benefits of data and shares them as widely as possible.</p>
<p>I'll be in touch.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>BBC+ relaunch - An experimental app for a more personal BBC</title>
      <description><![CDATA[Here's what's happening with the relaunch of the BBC+ app.]]></description>
      <pubDate>Wed, 12 Dec 2018 11:34:28 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/82cd8d1e-2f23-4eff-8f34-0ef38ca8854c</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/82cd8d1e-2f23-4eff-8f34-0ef38ca8854c</guid>
      <author>James Metcalfe</author>
      <dc:creator>James Metcalfe</dc:creator>
      <content:encoded><![CDATA[<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06vc28y.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06vc28y.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06vc28y.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06vc28y.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06vc28y.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06vc28y.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06vc28y.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06vc28y.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06vc28y.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>Open just about any mobile app and your experience is no doubt being tailored for you in some way by algorithms, personalisation, AI, Big Data, machine learning... any more buzzwords you can think of?</p>
<p>This can be great for us as users... we get a better feed, faster.</p>
<p>With the refreshed BBC+ app, we will be making recommendations using data to take steps towards re-inventing the BBC for a new generation.</p>
<p>It also brings a great opportunity to be more experimental in our approach. We will be testing ideas quickly to realise the possibilities of a data driven world both in the way we work and the way we make our products. We have some of the most popular apps in the UK, and know we could do a lot more to improve and make them more welcoming, relevant and personal. We also need to understand how, as a public service organisation, we maintain our values in a data driven world.</p>
<p>One of the biggest challenges with making our data more connected is that the information we have on our content is based on our organisational structure.This means you can indicate a topic that is relevant to you in the BBC News app (such as health) but that preference is not reflected in iPlayer and so we currently don&rsquo;t recommend all the great health shows you might like. This can be frustrating as you have to set up and manage each product separately.</p>
<p>To change this we are aiming for a more consistent and personalised experience. What if we could easily bring the best content we make, to you, from anywhere in the BBC? Would people discover and enjoy the BBC more? We need to take some risks, try some new and challenging ideas, and ask some tough questions as we work towards a more relevant and personal BBC for everyone.</p>
<p>With this in mind, we are today making an experimental update to our BBC+ app.</p>
<h4>Why change BBC+?</h4>
<p>BBC+ was set up as a personal BBC app with similar goal, however without the support of the data capabilities we are trying now. It didn't take off as we'd hoped. Much of what it currently offered is provided by BBC News and replicates the BBC Homepage.</p>
<p>In order to move forward we have decided to start again, while beginning to use machine learning and data. Our plan is to keep it small, to try new ideas as quickly as possible, learn as we go and improve based on regular feedback.</p>
<p>The new BBC+ app is a combined effort from the <a href="https://findouthow.datalab.rocks">Datalab</a> and BBC+ app development teams. We have also been working collaboratively with colleagues from across the BBC to understand the challenges we collectively face. My colleague Svetlana Videnova describes some of the <a href="http://www.bbc.co.uk/blogs/internet/entries/a26a25af-4012-4f00-9fe9-2cc639a76340">technical approach in her blog&nbsp;</a>and the Datalab team <a href="http://www.bbc.co.uk/blogs/internet/entries/50f047ef-d06b-487b-af3f-acfcb705c8e6">have released a free online course to help people understand machine learning, data science and how we work</a>.</p>
<p>For now the new BBC+ app contains short videos, however there is far more breadth and depth to explore than the old app. Our overall mission is to help you discover the gems that you didn&rsquo;t even know you were missing, at a time that suits you.</p>
<p>This is clearly an early starting point and as we learn more we expect to make changes often</p>
<h4>Why now?</h4>
<p>We are releasing this update at an early experimental phase as we believe using data and machine learning in the BBC is a shared challenge and needs to be done in a transparent way.</p>
<p>Working in a closed group limits how much we can learn about machine learning and so we want to invite your feedback and discussion throughout the process. We know there are improvements we need to make and we want people to be directly involved in shaping the next stage. Our team are keen to gather thoughts and ideas on improvements, and so please use the feedback buttons in the app to send in your thoughts and comments. The team will be reporting back regularly on progress!</p>
<p>BBC+ is available on Android: <a href="https://play.google.com/store/apps/details?id=uk.co.bbc.bbc_plus">https://play.google.com/store/apps/details?id=uk.co.bbc.bbc_plus </a></p>
<p>and iOS: <a href="https://itunes.apple.com/gb/app/bbc-the-bbc-just-for-you/id1110317391?mt=8">https://itunes.apple.com/gb/app/bbc-the-bbc-just-for-you/id1110317391?mt=8</a></p>
<p>&nbsp;</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06vc2f9.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06vc2f9.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06vc2f9.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06vc2f9.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06vc2f9.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06vc2f9.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06vc2f9.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06vc2f9.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06vc2f9.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Datalab representing machine learning in the BBC - the experiment</title>
      <description><![CDATA[How the BBC's Datalab is using machine learning to enhance our digital products and services.]]></description>
      <pubDate>Wed, 12 Dec 2018 10:06:17 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/a26a25af-4012-4f00-9fe9-2cc639a76340</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/a26a25af-4012-4f00-9fe9-2cc639a76340</guid>
      <author>Svetlana  Videnova</author>
      <dc:creator>Svetlana  Videnova</dc:creator>
      <content:encoded><![CDATA[<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06vbxnd.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06vbxnd.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06vbxnd.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06vbxnd.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06vbxnd.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06vbxnd.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06vbxnd.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06vbxnd.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06vbxnd.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>One of the key objectives for the BBC for the coming years is to focus on younger audiences.&nbsp;Machine Learning recommendation capabilities can help us achieve that.</p>
<p>In a recent speech to BBC staff Tony Hall described our shared ambition &ldquo;to grow our weekly online reach with younger audiences from 55% to 90% within four years&rdquo;. We expect that proving personalised content via the Datalab platform to the new BBC+ app, including the learning throughout the experiment, will begin to contribute to this growth.&nbsp;&nbsp;Product Manager James Metcalfe describes the approach and objectives of BBC+ in <a href="http://www.bbc.co.uk/blogs/internet/entries/82cd8d1e-2f23-4eff-8f34-0ef38ca8854c">his blog post</a>.</p>
<p>The aim of <a href="https://findouthow.datalab.rocks/">Datalab</a>, part of BBC Design + Engineering, is to help achieve this vision and reach younger audiences, via a new way of working and by experimenting with new <a href="https://www.youtube.com/watch?v=fz3YEX8NgtI">technologies</a>.</p>
<p>Datalab is the first BBC platform working with Google Cloud Platform <a href="https://medium.com/bbc-design-engineering/how-we-deliver-with-gcp-at-the-bbc-1c9812acf3a1">(GCP)</a> in production. Being the pioneer of this integration is meant to refine our approach to information security and the data privacy approval process, as well as establishing a new infrastructure.</p>
<p>As if this wasn&rsquo;t challenging enough, we brought more excitement to our engineering team by including <a href="https://www.youtube.com/watch?v=-weU0Zy4Yd8">gRPC</a> integration, Elasticsearch, Kubernetes and Spinnaker for container-based deployments and Drone as part of our stack. As a team the decision was made to implement these technologies, regardless of the fact that not everyone was experienced with them. As a result, we have adopted some more than others, but again, one of the key points is that we learn along the way, and gain skills that will be valuable later in the programme.</p>
<p>As a platform, we had to connect to the existing BBC data stores, including our User Activity Store (UAS) database, as well as serving the media content from a different AWS database.</p>
<p>This infrastructure allowed us to provide the groundwork for data scientists to start to explore various methods to satisfy the needs of BBC products and master the personalised experience for BBC users, starting with BBC+ app.</p>
<p>For example, new users, who have no previous history with the BBC, are given a cold start recommendation, so that they can begin their BBC+ journey. The first focus for the Datalab team was to create these recommendations to serve relevant content. It progressively incorporates a user&rsquo;s history, refines their content and helps them discover interesting BBC suggestions.</p>
<p>Finding out what data we had to work with was crucial. We quickly discovered that the metadata for some BBC content is inconsistent. This lead us to conversations with our editorial colleagues on ways of tagging and creating content metadatain a more uniform manner, so that we can surface their output to the audience in a more personalised way.</p>
<p>The first content type to be ingested is video clips. In the future the aim is to include audio and articles. Currently in our Elasticsearch DB we have 1,137,598 clips. A set of filters was applied to provide only relevant clips, with complete metadata and editorial risks mitigated:</p>
<ul>
<li>Unique clips</li>
<li>Only English content (for now)</li>
<li>Filter out audio and weather (for now)</li>
<li>Filter out clips older than 2013 (for now)</li>
<li>Editorial risk filtering</li>
<li>128 BBC brands are not surfaced in the BBC+ app</li>
<li>8 master brand were filtered out (mainly to help serve only English content for now)</li>
</ul>
<p>That leaves us to 131,626 clips available for BBC+ users.</p>
<p>The following techniques are being used by our data science team, to experiment, score and create &ldquo;better&rdquo; recommendations:</p>
<ul>
<li>Model-based collaborative filtering: we&rsquo;re using embeddings for our content using word2vecmodels, making the content -&ldquo;words&rdquo; and our playlists -&ldquo;sentences&rdquo;</li>
<li>Offline scoring: to measure how a particular recommender system is performing, we&rsquo;re using different metrics, like recency, popularity, Normalized Discounted Cumulative Gain (nDCG) and hit-rate. This helps us to select which versions of our models should go for online scoring</li>
<li>Online scoring: we have a system in place to score the recommenders with the live user data, using A/Bor hit-ratetesting, whenever necessary</li>
<li>Combining ML with editorial guidelines: we can prioritise one genre or brand against the others, to fit the editorial needs and better match our audience expectation</li>
<li>To help share what we know <a href="http://www.bbc.co.uk/blogs/internet/entries/50f047ef-d06b-487b-af3f-acfcb705c8e6">we created a course</a>, to establish a good understanding of what we are building</li>
</ul>
<p>We came out of this iterative process full of valuable insights, and building a new ML platform was just one of the things we discovered. Datalab established a positive team culture, with a great deal of multi-disciplinary learning. Despite the bumpy road, we now have a clear vision of our engineering and data science responsibilities, including A/B testing our process, trying various agile methodologies, and bringing teams in Salford and London together for closer collaboration.</p>
<p>We are excited about the next steps, and there is a lot to do on all aspects of our platform: infrastructure, engineering, devops, data science and ML. We will be inviting a new BBC team into the Datalab world: beginning with an exploratory session with the Voice Team over Christmas. We will also be proposing collaborations with other product groups, including R&amp;D and News, early in 2019 to continue to experiment, push the boundaries of our exploration further and innovate ML in the BBC.</p>
<p>If you would like to know more about Datalab, our journey, or you would like to share your own experience, feel free to contact us at datalab@bbc.co.uk.</p>
<p>And in other important news: we are still recruiting! <a href="https://findouthow.datalab.rocks/">https://findouthow.datalab.rocks/</a></p>
<p>BBC+ is available on Android: <a href="https://play.google.com/store/apps/details?id=uk.co.bbc.bbc_plus">https://play.google.com/store/apps/details?id=uk.co.bbc.bbc_plus</a></p>
<p>and iOS: <a href="https://itunes.apple.com/gb/app/bbc-the-bbc-just-for-you/id1110317391?mt=8">https://itunes.apple.com/gb/app/bbc-the-bbc-just-for-you/id1110317391?mt=8</a></p>
<p>&nbsp;</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>How we built BBC Sounds on the web</title>
      <description><![CDATA[We designed BBC Sounds as a completely new web application, re-evaluating architecture and taking advantage of new web technologies. It’s built upon our experience of developing and maintaining iPlayer Radio and BBC Music.]]></description>
      <pubDate>Mon, 26 Nov 2018 16:26:00 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/8baa1b0c-1f3f-405a-849d-7d19eb125d50</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/8baa1b0c-1f3f-405a-849d-7d19eb125d50</guid>
      <author>Julia Miroshina</author>
      <dc:creator>Julia Miroshina</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>In late 2017 engineering teams that had previously worked on iPlayer Radio and BBC Music began work on a new product. Launched 16th October 2018, <a href="https://www.bbc.co.uk/sounds">BBC Sounds</a> is an audio product from the BBC that brings together live and on-demand radio, music and podcasts into a single personalised experience.</p>
<p>In this blog post we&rsquo;ll talk about the technologies and engineering practices that helped us to build the BBC Sounds web application.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06swvc0.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06swvc0.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06swvc0.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06swvc0.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06swvc0.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06swvc0.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06swvc0.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06swvc0.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06swvc0.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>We designed BBC Sounds as a completely new web application. It&rsquo;s built upon experience from developing and maintaining iPlayer Radio and BBC Music. It allowed us to re-evaluate architecture and take advantage of new web technologies.</p>
</div>
<div class="component prose">
    <h4>Technology</h4>
<p><strong>Cloud infrastructure and performance</strong></p>
<p>The web version of BBC Sounds was released to the existing iPlayer Radio audience as well as absorbing traffic from other parts of the BBC web estate. It is a cloud hosted application built to scale to handle a large number of requests, with the business logic encapsulated into the API layer and shared amongst clients.</p>
<p>As our applications migrate into the cloud from on premise infrastructure our engineering teams assume DevOps responsibilities. We often use <a href="https://gatling.io">Gatling</a> for load testing, along with mock servers that provide upstream responses for specific scenarios. For BBC Sounds we developed new metrics and defined actionable alerting. This improved application availability making it more resilient and scalable than ever before.</p>
<p>To find out how BBC serves its web pages to the public, check out the post &ldquo;<a href="http://www.bbc.co.uk/blogs/internet/entries/328e1b75-26f9-49e9-9ed1-5abd481f03f3">How we deliver BBC pages to the internet</a>&rdquo;.</p>
<h4>New stack</h4>
<p>Building a new product provides an opportunity to start using new technology. BBC Sounds has been built using <a href="https://nodejs.org/en/">Node.js</a> and <a href="https://reactjs.org">React</a> with server-side rendering.</p>
<p><em>&ldquo;BBC Sounds was built from the ground up, using Node.js, React, Redux and Express. The migration to Node.js wasn&rsquo;t completely alien to us, as by 2017 we were already using it for client-side pipeline, such as Gulp, Sass, minifying and linting tasks. What was new was transitioning to server JavaScript using Express as web framework for Node.js. We chose it for its simplicity and minimal footprint, allowing us to get application setup quickly, but also providing engineers with freedom to define development style.&rdquo;</em><br />&mdash; Jason Williams, Principal Software Engineer, BBC Sounds</p>
<p>We began by thinking about the UI components, and how they would be laid out. Our existing Radio network home pages (Radio 1, Radio 2 etc) were also built from UI components, but they were completely independent packages separately fetched to build the page. The problem with this approach was that modifying the components in the browser remained difficult, as they were composed server-side (using <a href="https://twig.symfony.com">Twig</a> templates). We needed to change our approach.</p>
<p><strong>Building Components with Atomic Design</strong></p>
<p>By using React to build components we can render sections of the page server-side and then continue to modify them in the browser without duplication of code. Some of these components need to be reusable, so we needed a versatile, but consistent approach.</p>
<p>We used <a href="http://atomicdesign.bradfrost.com/chapter-2/">Atomic Design methodology</a> to break components up into distinct levels: atoms, molecules and organisms.</p>
<p>Here is an example using Radio 1&rsquo;s Listen Live roundel on the homepage:</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06swvk7.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06swvk7.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06swvk7.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06swvk7.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06swvk7.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06swvk7.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06swvk7.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06swvk7.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06swvk7.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Example of a component with atomic design</em></p></div>
<div class="component prose">
    <p>The process starts with the design of a component from our UX team. As you can see, they want to show the user what&rsquo;s playing live on our radio networks.</p>
<p>We start to break this down by identifying the smallest discrete units of the design e.g. the ProgressBar, the live Label, a ResponsiveImage etc. These are atoms which we can compose to represent a single network, lets call it a NetworkItem molecule. Finally we can represent our networks in a scrollable list &mdash; a Carousel organism.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06swvtk.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06swvtk.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06swvtk.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06swvtk.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06swvtk.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06swvtk.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06swvtk.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06swvtk.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06swvtk.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Example of NetworkItem being populated, data comes from API layer</em></p></div>
<div class="component prose">
    <p>Each component we then build and test separately, with API simplicity and reusability in mind. All that&rsquo;s left is to hydrate the components with the data from our API and voila, it renders! The more reusable components we build, the faster we can test different configurations with users, and deliver new sections of Sounds.</p>
<p><strong>Rendering on the server and isomorphism</strong></p>
<p>We run server-side JavaScript alongside a JavaScript front-end framework. It helps to avoid having disparate languages for rendering and interacting with the UI. The application renders the full markup of its initial state in the server context. Then, we deliver both the state and a copy of the transpiled application, to pick up where we left off client-side.</p>
<p>When a user visits a page in Sounds, the application responds with the static HTML for the page requested . Users don&rsquo;t have to wait for JavaScript to load. Then, as the React app mounts atop the existing HTML, components become more interactive; buttons appear on items to fetch more episodes from a podcast, or allowing users to subscribe.</p>
<p><em>"The benefits of this approach are multiple: ready-to-render initial responses are great for SEO and accessibility where good network connectivity and JavaScript is not always a given. But for engineers it&rsquo;s also powerful to remove the context switch between client/server languages and tools, to code one language everywhere."</em><br />&mdash; Stephen Roberts, Senior Software Engineer, BBC Sounds</p>
<p><strong>React</strong></p>
<p>Choosing a front-end framework from the plethora offered up by the JavaScript ecosystem is no small feat for a new product. Our decision to use React was ultimately guided by some core features we knew Sounds would need to be future-proof and flexible to changing requirements, but also scale horizontally with our development capability.</p>
<p>React is among the frameworks which implement a virtual DOM (VDOM). The VDOM is a layer of abstraction which represents the UI in memory and employs a &ldquo;<a href="https://reactjs.org/docs/reconciliation.html">diffing algorithm</a>&rdquo; to apply updates to the DOM efficiently given a change in application state. This frees us up from writing boilerplate, data-binding / DOM manipulation code, to build the audience-facing components which make BBC Sounds.</p>
<p>React is a mature, open-source framework from Facebook, with a large community of developers and learning resources. Many large-scale applications run on React in production externally to the BBC, but also internally &mdash; product teams have been reaping the benefits of using React and React-like frameworks. Choosing React therefore, leaves the door open for inter-product code and component sharing.</p>
<p><strong>Redux</strong></p>
<p>Delivering richer experiences around our content on the web translates to added complexity in frontend logic &mdash; so it becomes increasingly important for engineers to be able to: reason about the internal state of our JS, communicate state changes to UI components efficiently, and debug when things go wrong.</p>
<p>In Sounds, we use <a href="https://redux.js.org">Redux</a> as a predictable state container to provide our React app with a central, read-only store for its state. The state store can only be updated via a set of pure functions called reducers. These functions take the current app state plus a plain JavaScript object describing a state-change (an action), and update the store accordingly. React-redux then connects the data from the store to React components, creating an easy to understand pattern for unidirectional data flow.</p>
<p>In this way, application state is always predictable; any sequence of actions can be replayed on state with the same result. And, since the reducer functions that handle the changes are pure, they&rsquo;re very straightforward to test.</p>
<p>Just as we changed the way we build applications, we also changed the way engineering team works and adopted new practices.</p>
<h4>Engineering team practices</h4>
<p>Each software engineering team comprises of a project manager, product team (e.g. product owner, business analyst), embedded UX designers and QA testers, team lead and software engineers, with streamlined processes and agile methodology in place, similar to other BBC teams.</p>
<p>We use a variety of tools to speed up our development process and facilitate communication, e.g. GitHub, custom Slack integrations etc. There&rsquo;re also various practices that engineering team has been following &mdash; some of them are listed below.</p>
<p><strong>Cross-discipline collaboration</strong></p>
<p>Developers often pair with UX designers, especially for branding related tasks and fine-tuning interactive experiences. Developing the component library has also been a joint effort from both engineering and UX teams to ensure consistency for users.</p>
<p>Running regular sessions with engineers working on the service layer and collaborating with other client teams ensures consistency in requirements and reduces complexity for the clients. You can read more about <a href="http://www.bbc.co.uk/blogs/internet/entries/7e65dff6-9bf1-4f98-b185-3b424939c8ce">developing Sounds in the API</a>.</p>
<p><strong>Open source and inner source</strong></p>
<p>It has been an ongoing aspiration within the BBC to open source more as well as use inner source tools and libraries.</p>
<p>Here are examples of open-source BBC libraries that have been used for Sounds development on the web:</p>
<ul>
<li><a href="https://github.com/bbc/bbc-a11y">bbc-a11y</a> &mdash; BBC Accessibility Guidelines Checker</li>
<li><a href="https://github.com/bbc/express-simple-timing">express-simple-timing</a> &mdash; <a href="http://expressjs.com">Express</a> middleware that sets Server Timing API headers and optionally sends timers to stats systems.</li>
<li><a href="https://github.com/bbc/grandstand">BBC Grandstand</a> &mdash; collection of common CSS abstractions and utility helper classes</li>
<li><a href="https://github.com/bbc/speculate">speculate</a> &mdash; automatically generates an RPM Spec file for a Node.js project</li>
</ul>
<p><strong>Architecture decision records (ADRs)</strong></p>
<p>Following API-led product development, clients of the API contribute to architecture decisions. Depending on context and complexity, these decisions could be discussed in the form of pull requests, white-boarding sessions, and stubs.</p>
<p>Architecture design decisions are short documents which enable our teams to have a current understanding of the context and decisions the team is making. As decision is made, it gets documented as an ADR (<a href="http://thinkrelevance.com/blog/2011/11/15/documenting-architecture-decisions">Architecture Decision Record</a>) and added to version control with the code.</p>
<p><em>&ldquo;The hope is these can act as a unit of knowledge being passed across team boundaries to allow us, much like a distributed system made of people, to achieve an eventually consistent base of knowledge of our technical estate. They also form an historical record for a team, to help them bring people up to speed and remind themselves of their own technical foundations.&rdquo;</em><br />&mdash; Paul Caporn, Lead Technical Architect BBC TV&amp;Radio</p>
<p><strong>OKR methodology</strong></p>
<p>In Sounds each team is responsible for setting its own Objectives and Key Results &mdash; a methodology which is described in detail in the book &ldquo;Measure What Matters: How Google, Bono, and the Gates Foundation Rock the World with OKRs&rdquo; by J. Doerr. OKRs are meant to set strategy and goals over a specified amount of time for teams and individuals. At the end of that time period, OKRs provide a reference to evaluate whether these objectives have been met.</p>
<p><strong>A/B testing</strong></p>
<p>While developing Sounds we started extensively using A/B testing tools and success measures to ensure that the changes we&rsquo;re making are the right ones for our audience. Every visit of an audience member is an opportunity to improve the product. Our experimentation toolkit allows teams to quickly test variations against an existing experience to determine what the optimal approach is in terms of a set of metrics. You can find out more about experimentation in the BBC in this blog post on optimising iPlayer experience.</p>
<p><strong>10% time</strong></p>
<p>Sounds have been making use of 10% time initiative <a href="https://iplayer.engineering/autonomy-innovation-using-10-time-e04dc7f184e9">adopted by the BBC software engineering teams</a> that enabled developers to research and implement innovative solutions. Typical sprints in Sounds teams lasts two weeks, which means that one day per sprint is dedicated to 10% initiatives. This ensures 10% doesn&rsquo;t skew delivery estimates for project managers, allowing flat distribution of allocated time per sprint. 10% time has been highly beneficial for engineers, giving space to grow, learn and think outside the box.</p>
<h4>What&rsquo;s next?</h4>
<p>In 2019 we&rsquo;re looking to personalise BBC Sounds even more, and follow our data-driven approach to implement meaningful improvements. By knowing our audience we aim to deliver a better experience of sounds, innovate on content, challenge our users, and provide an equal opportunity to access the best content for everyone.</p>
<p>If you are interested in working on projects like BBC Sounds, solving exciting, complex and large scale problems, why not come and join us? <a href="https://careerssearch.bbc.co.uk/jobs/custom/?fields%5B32%5D=571">See latest jobs in D&amp;E TV &amp; Radio.</a></p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Building BBC Sounds in the API</title>
      <description><![CDATA[Patrick Cunningham explains some of the background architecture behind the new service BBC Sounds.]]></description>
      <pubDate>Thu, 08 Nov 2018 09:44:22 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/7e65dff6-9bf1-4f98-b185-3b424939c8ce</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/7e65dff6-9bf1-4f98-b185-3b424939c8ce</guid>
      <author>Patrick  Cunningham</author>
      <dc:creator>Patrick  Cunningham</dc:creator>
      <content:encoded><![CDATA[<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06r1r1g.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06r1r1g.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06r1r1g.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06r1r1g.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06r1r1g.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06r1r1g.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06r1r1g.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06r1r1g.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06r1r1g.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Radio and Music Services team (minus Dan and Ray)</em></p></div>
<div class="component prose">
    <p>For the past 3 years, the Radio and Music Services (RMS) team has been making APIs for iPlayer Radio, BBC Music and podcasts in the Radio and Music metadata space. We are an Agile Software Development team made up of 8 Software Engineers, 3 testers, a Product Owner, Project Manager and Business Analyst. Our vision is:</p>
<p><em>&ldquo;We encapsulate business logic for Radio &amp; Music products on all platforms. We add value by providing the right blend of metadata, reliably and fast&ldquo;</em></p>
<p>In 2017 the Radio and Music department were tasked with building a new personalised audio product built using an API first approach. I&rsquo;m going to talk about what that means to us and how we went about building Sounds in the API.</p>
<p>In the past typically we have made API endpoints designed for specific features our clients would request from us, they would write or modify existing client code to integrate with this and then deliver the feature. Clients would still integrate with various BBC APIs to build products but the benefits of having a platform independent API to provide a single integration point to these services were obvious. We decided for Sounds to go one step further and define the layout and content of the product in an API.</p>
<h4>The stack</h4>
<p>We love Scala (after we overcame the learning curve) and our APIs are built using predominantly <a href="https://doc.akka.io/docs/akka-http/current/introduction.html">Akka HTTP</a> and <a href="https://doc.akka.io/docs/akka/current/stream/stream-introduction.html">Akka Streams</a>. We&rsquo;ve found these tools provide excellent performance on modern cloud server architecture, and are scalable, resilient and a fantastic fit for the kind of concurrent retrieval of data we need from various BBC systems and databases.</p>
<p>Programme metadata is our bread and butter, and customising that to each users&rsquo; needs is at the core of Sounds. We have, at the time of writing, roughly 350,000 pieces of available audio we want users to be able to access. We think you should be able to serve up some of the oldest content, <a href="https://www.bbc.co.uk/sounds/play/p03yjdfq">Britain declares war on Germany</a> just the same as the latest <a href="https://www.bbc.co.uk/sounds/play/brand:b0080x5m/p00g3r1s">Breakfast show with Greg James</a>.</p>
<p>We follow the <a href="https://info.lightbend.com/COLL-20XX-Reactive-Microservices-Architecture-RES-LP.html">Reactive Microservices Architecture</a> made by the super smart people at <a href="https://www.lightbend.com">Lightbend</a> as much as we can.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06r1r9m.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06r1r9m.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06r1r9m.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06r1r9m.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06r1r9m.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06r1r9m.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06r1r9m.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06r1r9m.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06r1r9m.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Overview of Sounds API architecture</em></p></div>
<div class="component prose">
    <p>Do one thing and do it well.</p>
<p>We&rsquo;ve tried to stick to this principle to give us the flexibility to deploy and scale services individually.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06r1rj2.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06r1rj2.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06r1rj2.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06r1rj2.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06r1rj2.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06r1rj2.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06r1rj2.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06r1rj2.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06r1rj2.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>APIs — Do one thing and do it well, Database per microservice</em></p></div>
<div class="component prose">
    <p>The internal architecture at the BBC means that we typically get personalised information to hydrate programmes with from various sources:</p>
<ul>
<li>User Activity Service for play history, Bookmarks and Subscribing to Programmes</li>
<li>Recommendations API for programmes recommended to a user from their listening history</li>
<li>BBC Account for authentication to the BBC cross product Identification system</li>
</ul>
<p>We use the <a href="https://microservices.io/patterns/data/api-composition.html">API composition design pattern</a> for returning most of this content. The Personalised Programmes and Recommendations services are examples of this. We pass a users authentication token to an external service and get a list of programme identifiers back, then verify that those programmes are available and return a list of programmes back upstream.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06r27m1.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06r27m1.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06r27m1.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06r27m1.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06r27m1.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06r27m1.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06r27m1.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06r27m1.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06r27m1.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>API Composition in our Personalised APIs</em></p></div>
<div class="component prose">
    <p>A piece of work undertaken by a combination of Engineering, UX, Product and Project teams across RMS and client teams was to simplify the domain objects returned by the API. We hedged our bets that we would mostly be returning a combination of lists of programme content or single entities. We whittled it down to 4 main types:</p>
<p>1) Playable<br />The core of the product and an object you will see everywhere. In a nutshell, something you can play. It can be an episode, clip or live stream but clients don&rsquo;t need to know that</p>
<p>2) Display<br />A piece of content, image, link, text or placeholder</p>
<p>3) Container<br />A container of other programme objects. Typically this can be a programme brand or series, a category, radio network or editorially curated collection of programmes</p>
<p>4) Broadcast<br />Programmes with specific dates and times that are linked to Playable (on demands or live streams) or Display items (typically things that aren&rsquo;t available yet)</p>
<p>Each one of these object types was refined to return all necessary information but ONLY what was agreed with all teams. Making them as small as possible meant we optimised the size of content, efficiency and consistency of our responses.</p>
<p>We thought another good bet would be to define the experience of the product in the API and maintain a consistency across platforms. This also means there is one place to change layout, messaging and internationalisation. Clients are then free to apply their expertise in rendering that view in the best possible way for their platform. We began investigating returning all content required to load all the metadata content page in one request, benefitting mobile clients in particular. The continuing improvement of our services gave us more confidence in their performance and response times, that we could manage this all inside the Experience API and deliver all content apart from the platform specific layout on each. We started to develop this and found that not only could we improve page load times and latency, but client device and server performance, simplifying client code. This works particularly well for Search results, the Sounds homepage (Discover), Containers (Brand, Series and Collections) and My Sounds in the app.</p>
<p>It&rsquo;s been an enjoyable (but busy) year, with lots of interesting challenges and fun problems to solve. The team are really proud of their work and hope our users will get the benefit from it whether they&rsquo;re enjoying Sounds through the mobile apps (<a href="https://play.google.com/store/apps/details?id=com.bbc.sounds&amp;hl=en_GB">Android</a>, i<a href="https://itunes.apple.com/gb/app/bbc-sounds/id1380676511?mt=8">OS</a>) or the <a href="https://www.bbc.co.uk/sounds">web</a>.</p>
<p>In 2019 we&rsquo;re looking to personalise BBC Sounds even more through audience segmentation and releasing new features through multivariate testing, providing true cross platform Multivariate tests and measurement capabilities to our clients. We&rsquo;re also looking at content discovery feeds such as popularity allowing the audience to find more of the audio content they love and improving the &lsquo;continuous play&rsquo; experience. We&rsquo;ll let you know how we get on.</p>
<p>If you are interested in working on projects like BBC Sounds or a wider variety of exciting, complex and large scale problems, come and join us: Latest jobs in <a href="https://careerssearch.bbc.co.uk/jobs/custom/?fields%5B32%5D=571">D&amp;E TV &amp; Radio</a>.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Opening up our data science course materials</title>
      <description><![CDATA[Felix Mercer Moss explains how the Datalab course materials are being opened up to the public.]]></description>
      <pubDate>Tue, 23 Oct 2018 11:30:00 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/50f047ef-d06b-487b-af3f-acfcb705c8e6</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/50f047ef-d06b-487b-af3f-acfcb705c8e6</guid>
      <author>Felix Mercer Moss</author>
      <dc:creator>Felix Mercer Moss</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>Today, the <a href="https://findouthow.datalab.rocks/">BBC Datalab team</a> is releasing onto the open web the <a href="https://github.com/bbc/datalab-ml-training">course materials</a> it developed for newcomers to data science at the BBC.</p>
<p>Data science, and particularly machine learning, is trending like never before. At just 12 minutes (!), <a href="https://en.wikipedia.org/wiki/Conference_on_Neural_Information_Processing_Systems">NIPS tickets</a> <a href="https://www.forbes.com/sites/williamfalcon/2018/09/05/the-new-burning-man-the-ai-conference-that-sold-out-in-12-minutes/#359ec6087a96">sold out quicker</a> than <a href="https://en.wikipedia.org/wiki/Burning_Man">Burning Man</a> this year and if anyone still needs convincing, check out <a href="https://trends.google.com/trends/explore?date=2004-01-01%202018-10-15&amp;geo=GB&amp;q=data%20science,machine%20learning,hipster">this plot</a>:</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06pmhyj.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06pmhyj.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06pmhyj.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06pmhyj.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06pmhyj.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06pmhyj.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06pmhyj.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06pmhyj.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06pmhyj.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>It shows us that as recently as this year, both terms overtook the word &ldquo;hipster&rdquo; in web searches. While this may not be great news for <a href="https://www.bloomberg.com/news/articles/2017-05-03/from-unicorns-to-avocado-toast-hipster-fads-jack-up-food-prices">avocado sales</a>, it does look like it&rsquo;s a good time to be someone working with data! But as ever, with exposure can often come <a href="https://www.kdnuggets.com/2018/07/cartoon-data-science-religion.html">confusion</a>. So first, let&rsquo;s try and get some definitions straight.</p>
<p>The BBC handles terabytes (read - &ldquo;a lot&rdquo;) of new data every day, and data science is the field of study that helps us use that data to make better decisions and deliver greater value to our audiences. Machine learning, meanwhile, is just one class of statistical techniques widely used in data science that, as it turns out, is particularly powerful. However, it is not magic. Instead it is a collection of tools and techniques that use mathematics to draw conclusions from data.</p>
<p>It could be argued that nothing indicates more the need to publish a training course than the moment a topic surpasses in web traffic a cool word millennials (<a href="https://www.telegraph.co.uk/travel/advice/the-era-of-the-millennial-hipster-is-almost-over/">usedto</a>?) use. But our team did have other reasons for opening this course up. Here are just three:</p>
<p>1. To get people excited about learning more data science and machine learning; particularly when <a href="https://www.linkedin.com/pulse/skills-companies-need-most-2018-courses-get-them-paul-petrone/?trk=li_corpblog_jobs_skills_2018">data literacy is so high on many employers&rsquo; wish lists. </a></p>
<p>2. To share some of the interesting problems faced by the BBC in the data space.</p>
<p>3. To demonstrate how large organisations such as the BBC can use their audience&rsquo;s data to generate a positive impact.</p>
<p>Continuing to create engaging content, while the expectations of our (especially younger) audience members are <a href="https://qz.com/1128973/people-watch-netflix-at-work-and-in-public-bathrooms/">constantly shifting</a>, is one of our greatest challenges. To ensure we meet these new expectations, it is important for us to analyse and understand the conditions and behaviour patterns that lead to more, or less engagement.</p>
<p>In this course, we use data to explore the question: &ldquo;What makes BBC audiences engaged?&rdquo;. If you have at least a basic understanding of Python programming (statistics would be a bonus!) and a healthy interest in taking your first steps in data science, we think you will have a lot of fun exploring our course.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06pmk35.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06pmk35.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06pmk35.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06pmk35.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06pmk35.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06pmk35.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06pmk35.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06pmk35.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06pmk35.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>And if you found this training easy and had fun doing it, why not join us?</em></p></div>
<div class="component prose">
    <p>We take readers on a four-part journey that follows the typical pattern of many data-led projects. Each part is expected to take readers around one hour to complete and focuses upon: <em>data exploration, data transformation, classification models and regression models. </em></p>
<p>In <em>data exploration</em>, we first look into how to formulate our data science problem and perform some preliminary analyses to get a better feel for our data. In <em>data transformation</em> we then introduce readers to some basic machine-learning theory and walk through the process of how we prepare our data for ingestion into our statistical models. In the final two parts (<em>classification</em> and <em>regression</em>), the actual machine-learning starts, where we explain how to train, evaluate and choose the most appropriate model for our purpose.</p>
<p>The dataset that you get to work on contains the logs from 10,000 BBC iPlayer users. As you might expect, a public service broadcaster like the Beeb doesn&rsquo;t take data privacy lightly. So while the dataset we use is &lsquo;real&rsquo;, you can be sure that we have enforced particularly strong anonymisation so that the identity of the users is impossible to recover.</p>
<p>With an introductory course like this, we don&rsquo;t expect to make anyone an expert in data science overnight. However, we do hope that those of you who do take it will have a lot of fun, while also gaining some valuable insight into the decisions we make when working with data at the BBC. The topics covered in the online course only scratch the surface of the data science problems we encounter daily at the BBC.</p>
<p>If you would like to find out more about the challenges we face and how we are using data science and machine learning to find innovative solutions to connect with audiences, please get in touch!</p>
<p>We are always looking to improve the content of the course, so if you have any feedback or ideas for further instalments we would really like to hear from you.</p>
<p>Link to course: <a href="https://github.com/bbc/datalab-ml-training">https://github.com/bbc/datalab-ml-training</a></p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>My Life, My Data, #MyTomorrow</title>
      <description><![CDATA[Highlights from the Tomorrow's World campaign around data privacy.]]></description>
      <pubDate>Thu, 24 May 2018 14:17:03 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/969b8e44-cec7-40ef-ab39-c63d710f890b</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/969b8e44-cec7-40ef-ab39-c63d710f890b</guid>
      <author>Chris Sizemore</author>
      <dc:creator>Chris Sizemore</dc:creator>
      <content:encoded><![CDATA[<div class="component prose">
    <p>Tomorrow(&rsquo;s World) starts today.</p>
<p>Personal data is one of the most important issues we&rsquo;re all grappling with these days, but it can all feel so confusing and abstract that we tend to dismiss it with a mixture of 😕, 🤯,&nbsp;&macr;\_(ツ)_/&macr;.</p>
<p>What is personal data anyhow? It&rsquo;s data about you, but is it your friend, or your enemy? Often when it comes to personal data, a sense of dystopia can understandably prevail - we&rsquo;re being spied on, manipulated, and our civil liberties are under threat. Meanwhile, services that collect data about us have become so useful and seemingly indispensable (and those Buzzfeed quizzes ain&rsquo;t gonna do themselves).</p>
<p>In a world where it can often feel that things are constantly happening <strong>to</strong> us, how can we influence our own futures in a positive, active way?</p>
<p>It&rsquo;s in this context that the General Data Protection Regulation (GDPR) kicks into effect this Friday, 25 May 2018 - and it&rsquo;s a game-changer for people living in the UK and Europe when it comes to their rights to privacy in the digital age and what they can actively do with data about themselves.</p>
<p>To help audiences explore the implications of this monumental intervention, Tomorrow&rsquo;s World is launching &ldquo;My Life, My Data, #MyTomorrow&rdquo;, a campaign about people and communities shaping the future by taking control of the data they create. The campaign starts today and runs for the following week.</p>
<p>Using short films, interactive experiences, and conversations across social media, Tomorrow's World will help explain just how exciting and important personal data is.</p>
<p>Highlights of the &ldquo;<strong>My Life, My Data, #MyTomorrow</strong>&rdquo; campaign include:</p>
<p><strong>&ldquo;Future Values, or A Short Ride in an Intelligent Machine&rdquo;&nbsp;&nbsp;</strong>Data about you is helping drive the development of artificial intelligence. &ldquo;Future Values, or A Short Ride in an Intelligent Machine&rdquo; which launches today on <a href="https://www.bbc.co.uk/taster/">BBC Taster</a>, will send you into a &ldquo;what if?&rdquo; future built atop your own data. You&rsquo;ll exchange banter with the artificial intelligence behind a driverless taxi, and discover the guiding principles &nbsp;that make up <a href="https://www.bbc.co.uk/taster/pilots/global-values-where-do-you-fit">your own deepest values</a>.&nbsp;&nbsp;Experience <a href="https://www.bbc.co.uk/taster/pilots/tw-driverless-car">&ldquo;Future Values, or A Short Ride in an Intelligent Machine</a>". <a href="https://charisma.ai/introduction">Read more about Charisma.ai</a>, the cutting edge conversational storytelling technology behind this pilot.&nbsp;</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p0687sgt.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p0687sgt.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p0687sgt.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p0687sgt.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p0687sgt.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p0687sgt.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p0687sgt.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p0687sgt.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p0687sgt.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p><strong>&ldquo;My Life, My Data&rdquo;&nbsp;</strong>A short film (5 minutes) hosted by Leila Johnston and Alex Lathbridge takes us on a fast-paced journey to explore what is personal data and how is it currently used. Brought to life with animations by Jamie Squire, the short film reminds us of the quid pro quo that comes with access digital services such as Facebook, Google and Instagram: nothing comes for free.&nbsp;<a href="http://www.bbc.co.uk/guides/zjkxh39">View here</a>.</p>
<p><strong>&ldquo;Instagram Chatbot&rdquo;&nbsp;</strong>To illustrate the power of data about you and how it is driving the development of machine learning, our &ldquo;Instagram Chatbot" will analyse your next Instagram post and compare it with thousands of others to estimate how people will react to it. The chatbot will then create a unique, personalised short video explaining why and to illustrate the power of personal data. <a href="http://www.bbc.co.uk/guides/z7x9dxs">Try it here</a>.</p>
<p><strong>&ldquo;Donate Your Data Day&rdquo;&nbsp;</strong>A short film (10 minutes) that imagines a not-to-distant future where we can donate data about ourselves en masse, using a click-to-donate button on our mobile phones. Galvanised around a gently satirical global 'Donate Your Data Day', the audience can decide what data to donate and who to donate it to, showing their support by becoming a data donor with their own Data Donor Card. <a href="http://www.bbc.co.uk/guides/zrh347h">View here</a>.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p0683lyb.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p0683lyb.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p0683lyb.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p0683lyb.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p0683lyb.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p0683lyb.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p0683lyb.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p0683lyb.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p0683lyb.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""><p><em>Tomorrows World presenters Leila Johnston and Alex Lathbridge</em></p></div>
<div class="component prose">
    <p><strong>&ldquo;Meet the Personal Data Superheroes&rdquo;&nbsp;</strong>A special episode of the Tomorrow&rsquo;s World podcast explores the work of people making a difference when it comes to our data rights. Meanwhile, on social media, BBC presenters, journalists, and influencers - including Radio 1 Newsbeat&rsquo;s Tina Daheley, CBBC&rsquo;s Katie Thistleton and BBC Click&rsquo;s Spencer Kelly -&nbsp;discuss the importance of the data we're creating everyday. <a href="https://www.bbc.co.uk/programmes/p06830vl">Listen here.</a>&nbsp;</p>
<p><strong>"How Do You Feel About Your Data? A Survey"</strong> - a timely research project from Tomorrow's World partners The Open University, on their new nQuire citizen science platform. How much personal information are you happy to share? What should companies be allowed to do with the data you create? Audiences can contribute to this short survey and give their views on data protection. <a href="https://nquire.org.uk/mission/my-tomorrow">View here</a>.</p>
<p>The &ldquo;<strong>My Life, My Data, #MyTomorrow</strong>&rdquo; campaign will make you smile and think, and ask you to reconsider your preconceptions about the relationship between us, the data we create, and the companies, governments, and other organisations that use that data.</p>
<p>So what comes next?</p>
<p>This is just the beginning, and there's much more to do. People are recognising their rights and starting to take greater control of the data they create. They are actively helping create a future their grandchildren will want to live in. We can all join in.</p>
<p>Together, we're creating <a href="https://twitter.com/hashtag/MyTomorrow?src=hash%22%3E">#MyTomorrow</a>.</p>
<p><a href="http://www.bbc.co.uk/guides/zcxb7p3">Read more</a>&nbsp;about the Tomorrow&rsquo;s World partnership between Science Museum Group, Wellcome, The Open University, the Royal Society, and the BBC.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
    <item>
      <title>Your data matters</title>
      <description><![CDATA[How the BBC is meeting its GDPR requirements.]]></description>
      <pubDate>Tue, 22 May 2018 14:04:00 +0000</pubDate>
      <link>https://www.bbc.co.uk/blogs/internet/entries/7c605523-8df3-4dcb-bf58-7c64aa0b59a5</link>
      <guid>https://www.bbc.co.uk/blogs/internet/entries/7c605523-8df3-4dcb-bf58-7c64aa0b59a5</guid>
      <author>Julie Foster</author>
      <dc:creator>Julie Foster</dc:creator>
      <content:encoded><![CDATA[<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p06831gj.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p06831gj.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p06831gj.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p06831gj.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p06831gj.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p06831gj.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p06831gj.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p06831gj.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p06831gj.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>Last <a href="http://www.bbc.co.uk/blogs/aboutthebbc/entries/77bdafd0-20b3-414d-aa53-48786b194543">May</a>, we updated everyone on our plans to make the BBC more personalised and relevant to you. We can give you more of what you love when we understand you better, and also make sure that as a public service, we make something for everyone.</p>
<p>We now have over 15 million people with BBC accounts using the BBC&rsquo;s websites and apps in the last month. What&rsquo;s more, they are also using BBC websites and apps more than people who are not signed in. 64% of BBC account users visit BBC online more than 2 days per week, compared to 46% of all users. And when they are on BBC websites and apps, people with BBC account spend an additional hour per week than people not signed in.</p>
<p>Your personal data is helping power this transformation. We can&rsquo;t provide you with a meaningful personal or tailored experience without this information, but it is ultimately your data. And your data matters.</p>
<p>The General Data Protection Regulation, or GDPR for short, is coming into enforcement in the next week. It makes sure that businesses clearly explain to you why they collect your personal data, and how they use it. It is an evolution of the Data Protection Act, and gives you new and important rights.</p>
<p>As we&rsquo;ve said before, we&rsquo;ve built our new BBC account system with GDPR in mind, but we&rsquo;re always reviewing our processes, technology and governance.</p>
<p>We use your personal data for different reasons, and it&rsquo;s important we are transparent to you why we collect and use this data. Our site <a href="http://www.bbc.co.uk/usingthebbc">&lsquo;Using The BBC&rsquo;</a> spells out, in plain English, what we will (and importantly won&rsquo;t) do with your data. It also can help you exercise your GDPR rights, such as changing some of your details in Settings.</p>
<h4>How have we prepared for GDPR?</h4>
<p>For starters, you should not need to be a rocket scientist to know your rights. We&rsquo;ve updated our policies to make them even more transparent and clear.</p>
<h4>What are my rights?</h4>
<p>We&rsquo;ve created a new section in Using The BBC all about GDPR to help you understand what your rights are, how you can exercise them with the BBC and get help.<br />We&rsquo;ve also innovated and developed technology with data privacy at the heart of what we create. Below are a couple of examples of the kind of work we&rsquo;ve done to prepare for GDPR.</p>
<h4>Privacy for children</h4>
<p>We want to help you make sure that your child can only watch programmes, read comments and upload their creations in a space that is age appropriate and suitable to them. For this reason, we&rsquo;ve developed a way for parents or guardians to register their child which is simple and easy to do, but more importantly is safe and secure for the entire family. We want your children to get great experiences on BBC websites and apps, and play our part to help protect them.</p>
<h4>Privacy by design</h4>
<p>This little button means a lot to us.</p>
</div>
<div class="component">
    <img class="image" src="https://ichef.bbci.co.uk/images/ic/320xn/p067w1hg.jpg" srcset="https://ichef.bbci.co.uk/images/ic/80xn/p067w1hg.jpg 80w, https://ichef.bbci.co.uk/images/ic/160xn/p067w1hg.jpg 160w, https://ichef.bbci.co.uk/images/ic/320xn/p067w1hg.jpg 320w, https://ichef.bbci.co.uk/images/ic/480xn/p067w1hg.jpg 480w, https://ichef.bbci.co.uk/images/ic/640xn/p067w1hg.jpg 640w, https://ichef.bbci.co.uk/images/ic/768xn/p067w1hg.jpg 768w, https://ichef.bbci.co.uk/images/ic/896xn/p067w1hg.jpg 896w, https://ichef.bbci.co.uk/images/ic/1008xn/p067w1hg.jpg 1008w" sizes="(min-width: 63em) 613px, (min-width: 48.125em) 66.666666666667vw, 100vw" alt=""></div>
<div class="component prose">
    <p>We really want you to have a personalised experience, like picking up where you left off watching a show, getting recommendations on programmes you might like or getting notifications about your favourite football team. But you have the right to <a href="https://account.bbc.com/account/settings/privacy">turn off</a> these features if you don&rsquo;t want them. Our analytic services team has worked hard to develop a technical solution that can do this easily, and ensure your privacy.</p>
<h4>What&rsquo;s next?</h4>
<p>We have some fantastic events coming this summer, from the <a href="http://www.bbc.co.uk/mediacentre/latestnews/2018/biggest-weekend-broadcast">Biggest Weekend</a> to <a href="https://www.bbc.co.uk/sport/football/44103384">FIFA World Cup </a>&nbsp;and Wimbledon. Your data is helping us learn what you like, so we can make sure you get the best out of this summer, and improve our services for you in the future.</p>
</div>
]]></content:encoded>
      <slash:comments>0</slash:comments>
    </item>
  </channel>
</rss>
