I'm not sure about my guess that there would be 1Gb/s broadband by 2012 - but the rest of it seems solid. The thing is, there's still nearly three years to go. The rate of change is so great that I would actually be less confident about making predictions now than I was four years ago.
Bluesky predictions for the 2012 Olympics - from May 2004
Freelance technical journalists develop some extraordinary skills, probably the most important of which is staring out of the window for hours on end instead of writing a chronically overdue article. Inevitably, as the deadline approaches, miraculous things happen. The ironing gets done, and the house looks spotless. Any task – the more tedious or obnoxious the better – is preferable to the prospect of sitting in front of a blank screen and bashing out a couple of thousand words on the designated subject.
This time, though, staring out of the window at the (infrequently) blue sky was a necessary antecedent to the production of this article, because what you’re about to read is almost entirely speculative, and looks much further ahead into the future than anyone should reasonably feel comfortable with. Seven years, to be specific.
I’m referring, of course to the London Olympics (or, to be completely politically correct the “British” Olympics).
From a UK perspective, this is going to be a big event. I’ve heard that it’s going to involve the biggest civil engineering project ever. When you consider that it may even make it feasible to get from the Elephant to Lewisham at more than six miles an hour, you begin to realise the magnitude of this endeavour. And it’s probably reasonable to expect that it’s going to be the biggest broadcasting event ever, too.
So, now that we know that the Olympics is coming to London, it seems like a good time to speculate about the shape and form that such a broadcast event might take. I don’t have any insider information on this (I’m not sure if anyone does at the moment) but we can at least have a stab at it. And probably the best place to start is to look at the trends already hinted at by the phenomenon called “convergence”.
If you’ve blinked at any time recently you’ve probably missed it. Convergence has happened already. IT and telecommunications are now integral to acquisition, production and broadcasting, and it’s a very long time since anyone watched a television program that hadn’t been in and out of the digital domain at some point in its lifecycle. Computers and digital communication are an inescapable part of most people’s daily routine. I’m probably not completely typical, but, today, for example, I’ve watched the live output from BBC News 24 on my laptop via a 3G datacard, and talked to colleagues in Dubai and America using Skype, the popular (and free) Voice Over Internet Protocol phone service.
While convergence trends have been discernable for several years, it’s only recently that the benefits have been available to the masses, and the rate of progress is accelerating.
There are several reasons for this. First, virtually every aspect of computing technology is getting better, faster. Processors are effectively a thousand times quicker than they were around twelve years ago. Both RAM and hard disk storage are at least three orders of magnitude cheaper as well. Harder to quantify, but just as important, is that our knowledge of how to work with digital media has grown, as well as the sophistication of software tools and our ability to compress and process video and audio.
Even if all of these factors were (merely) developing linearly, the compound effect would be an acceleration of capability. In reality, the rate of change is even faster. So much so, that it’s getting harder and harder to look even a couple of years into the future. Ironically, having to stick with existing technological standards is helping us to look ahead with a little more clarity because they have a damping influence on our ability to innovate (DAB radio in the UK is a good example of this. It uses a technology devised in the 80s that is agonisingly out of date in comparison with the compression and digital transmission techniques available today).
There isn’t the space here to invent the entire future of broadcasting, although I’m sure we could all speculate endlessly. Instead, I’m going to do a rapid “flypast” through the areas where I think we’ll see the biggest change in the way sporting events will be broadcast
Resolution
This is an easy one. The London Olympics will be broadcast in high definition. So will the Bejing games, as well as next year’s world cup. What will be different, though, with the London event is that virtually everyone will be watching in high definition. It’s probably going to take seven years from now for the majority of televisions to be updated.
But, although it’s virtually certain that the games will have “better than PAL” coverage, it’s definitely not clear what the characteristics of the acquisition format will be.
For a start, there’s no reason to expect that video cameras will be restricted to any of the current HD raster sizes (1280 by 720 and 1920 by 1080). What’s more likely is that acquisition and production resolutions will be completely decoupled from delivery resolutions. It’s easy to derive virtually any raster size from a higher resolution (especially if progressive formats are used throughout), so there’s every possibility that we’ll see camcorders with 4K resolution or above. This might seem unrealistic, but you can already buy digital still cameras (such as the Canon Ds1 Mk II) with a sixteen megapixel resolution: arguably better than film. Of course you then have to deal with the torrents of data: a single frame at this size is 50 Megabytes!
What would be the point of shooting in Ultra HD resolutions? At the risk of stating the obvious, the pictures are better. What’s not so obvious is that the pictures are better whatever resolution you’re viewing them at, because they started out with so much information. The better the original images, the better job compression algorithms can do with the material. So, although this is probably more an academic point than anything else, shooting in ultra high definition could even improve the picture on your mobile phone.
For sport, there are more potential benefits when the acquisition format exceeds the resolution of the delivery format.
If the pictures from a video camera are, say, four times the resolution of the pictures that are broadcast, then it’s possible to zoom in by a factor of four without any apparent loss of quality. This gives enormous flexibility to “post produce” the event. It’s effectively like having more cameras and more operators.
Once you start down this path, it takes you a very long way indeed.
In still photography, you can use software applications to “stitch” together several overlapping images. It’s a process that works surprisingly well, and can yield impressive panoramic shots or whole patchwork quilts of landscapes that have a quite astonishing resolution. (Google Earth is perhaps the ultimate embodiment of this technique).
There’s really no difference between video and still photography, except that video effectively takes a lot of photographs in a given time. So there’s no reason why the output from several suitably arranged cameras shouldn’t be stitched together, in real time, to form a giant, contiguous, moving vista, which a sports editor could then pan and zoom around, effectively framing his or her own shots.
As the software improves, even the output from cameras that haven’t been deliberately aligned can be incorporated.
(This, incidentally, has great potential for security applications, because it would give the police, for example, the ability to amalgamate the output from all the cameras at a venue, and then perform searches based on timecode and the possible location within a stadium of the suspect).
Perhaps the ultimate manifestation of this technique is the ability to invoke “virtual” camera positions.
Here’s how it would work.
It’s already possible to create 3D landscapes using the output from two cameras. Most of us have seen 3D flypasts of the surface of Mars, generated from two cameras shooting the same scene from a slightly different viewpoint.
A sports stadium is a much more controlled environment, and it may just be possible to arrange a complete array of cameras, suitably spaced, so that by analysing and concatenating the combined output it would be possible to generate a speculative “virtual” viewpoint.
This is hard enough to do with still images, never mind with moving video, in real-time. But the last ten years has shown us that if you can think about something, without coming across contradictions or sheer impossibilities, then it’s probably going to happen, given merely enough engineering skill and processing power.
Analysis
Do you remember the Channel 4 Snickometer? In case you don’t, it was a gadget invoked during live cricket matches to give viewers a visual display of the sounds from an LBW incident, synchronised to slow motion video playback of the event. This quasi-scientific approach has now been joined by Hawkeye, which uses multiple camera positions to predict the trajectory of balls.
There’s almost no limit to the amount of analysis and resultant viewer aids. It’s all a question of computing power, and, of course, there being a sensible balance between gimmickry and genuine usefulness.
Ultimately, we’re going to be able to create real-time, detailed dynamic models of sporting action of virtually any complexity. We’ll be able to apply highly evolved heuristics (common-sense or experience-derived rules), based on years of sporting analysis, to live action. We’ll be able to create entire teams of virtual footballers and athletes. A lot of the technology is here today.
You only have to look at next-generation games consoles to see how we might get there sooner than you’d expect. The clue is that these devices have not only superb graphics capabilities, but astonishing real-time “physics” engines. So, what might look like a graphical facsimile of a boxer, will actually behave like a boxer, right down to the way the virtual bones, muscles and skin react on receipt of the winning punch.
So I think we can expect to see action replays that superimpose real-time anatomical analyses. And of course masses of data to accompany the pictures.
Metadata
On the face of it, Metadata is not a sexy subject. But just wait, because metadata and video are ultimately going to become the same thing.
If we are in a position to create virtual models of reality, using advanced motion-capture techniques that don’t really on athletes wearing ping pong balls on their elbows, then, as the technology improves, the models will ultimately be indistinguishable from the real thing. At that point, we leave pixels behind and transmit video using metadata alone. (If you’ve ever used Adobe Illustrator or Corel Draw, you’ll understand what I mean by this). We’ll be able to broadcast great-looking video to any device, which will be displayed at the maximum resolution of the screen. There’ll be perfect slow-motion as well, because this technique is temporally resolution-independent as well as spacially. This may be something that doesn’t happen until way beyond the London Olympics, but, even so, metadata’s going to have a huge role in future sporting action.
The more metadata you can generate automatically, the more value your video has to anyone searching for it. An automatically generated metadata tagging schema, based on a standard taxonomy (a taxonomy is a hierarchy of categories, like Animals/Mammals/Primates/Chimpanzees etc) will allow productive searches by anyone who’s looking for highlights in their favourite sport. Comprehensive automatically generated metadata will help create automatically formatted web, video-rich websites, which can be completely tailored to an individual’s preferences. You can think of this in an abstract sense as giving the viewer control over the playout server’s schedules!
Bandwidth
By 2012, bandwidth will be as abundant as mains electricity. We’ll no more worry about bandwidth constraints than we would about plugging a mobile phone charger into a mains socket. And that includes mobile devices. The first 24 megabit/second DSL service has just been announced in London. In seven years time, we could have gigabit into the home, and, at last, mobile phones will show decent video.
I think it’s likely that download-store-play will replace standard, linear broadcast viewing habits. With rich metadata, video podcasting and semantically enabled Tivos, changing channels will be a meaningless activity, to be replaced by merely expressing the nature and strength of your viewing preferences.
Of course, there will always be live broadcasts. And, although there will always be the option to be “spoon fed” with the directors take on the event, viewers will be able to select camera angles, action replays and even move a “virtual” camera to any position imaginable.
DRM
Every technology has a “damping factor” that slows down its adoption. And, as we’ve seen above with legacy broadcasting standards, they can act both as an enabling and disabling influence. With major sporting events, this factor may well be Digital Rights Management.
No one imagines that organising the greatest broadcasting event ever seen will be entirely without cost. The political and economic wranglings that surround sport are legendary and are likely to be exceeded by orders of magnitude where the London Olympics are concerned.
The combination of advanced interactive acquisition technology with the internet is a DRM nightmare. Have you ever tried to listen to Radio 5’s coverage of football matches on the internet when you’re abroad? More often than not, they’re blanked out; presumably because of rights issues.
The only way to walk this particular tightrope is to make use of detailed metadata to derive permissions for who can watch what from where. It’s a rather dry subject, superficially devoid of interest, but could scupper whole swathes of new-generation media coverage.
So, is all of the above prescient, visionary reporting, or is it third rate science fiction?
I’ll let you know in seven years.