BlueFin2Last week, in the run-up to his sixth birthday, my son asked me an awkward question – what shape is the internet? With a gift for finding the existential fault lines between the conceptual and the tangible, his questions regularly leave me dithering and exasperated. But the question is an intriguing one and deserves thought...

Rapidly running out of planes and dimensions to imagine in, the only conclusion that I reached is that whatever shape it is, it is certainly, in some way, exponential. Everything about the internet is exponential – the growth of number of hosts, domains and Facebook accounts, the amount of traffic – all has grown exponentially. Geometric progression seems to be woven into the fabric of the Internet. Why?

Certainly demand is a factor – as a species we love to communicate in tiny misspelled messages and have an unlimited appetite for watching gorillas play the drums – but this is only a part of the story. It is no accident that internet expansion mirrors the growth of the technology that underpins it. As the processing and communication capacity of the vast interconnected network of fibres, wires and silicon humming and whirring in climate controlled rooms across the world doubles, so the number of emails, tweets and gorillas doubles. This growth is both tracked and predicted in the supremely elegant Moore’s Law which describes the trend in integrated circuit fabrication in which the density of transistors doubles every two years.

Moore's Law

This claim first appeared in a 1965 paper written by Gordon E Moore, but it was not until 1970 that Caltech professor Carver Mead termed the phrase Moore’s Law. It is often misquoted, variously being misapplied to processing power and memory density, and predictions of its demise are a more or less annual event. But despite this, the law continues to prove unswervingly accurate.

Clearly, Moore’s Law is not an actual law, at least not in the sense of the laws of physics, which are eternally inviolable, or in the legislative sense, where on-the-spot fines might be handed to semiconductor manufacturers failing to meet the mandated geometric feature density. Rather, it is simply a prediction, albeit a fabulous one – a shining gem set in a sea of awful predictions. There is an alternate view that it has become self-fulfilling prophecy as the law sets the bench-mark for technological improvement. This gives rise to an interesting thought – if Moore had, in 1965, claimed that recent impressive rate of improvements would just fizzle out because computers were already powerful enough, then would today’s mobile phones be the size of a garden shed?

The headline beneficiary of Moore’s Law has been microprocessors. The very first of these, the Intel 4004, contained around 2,300 transistors. Today’s Quad-core Itanium packs two billion. This information allows us to put Moore to the test. The 4004 was introduced on 15 November 1971. The Quad-core Itanium was released 8 February 2010, about 38.6 years later. Applying Moore, 2,300 x 2(38.6/2) = 1.48×109 or about 1.5bn. It seems that Moore’s law is being exceeded, as, by this calculation, Intel should hit two billion 19.73 years after the 4004, or April 2011 – not too far ahead of schedule in the scheme of things. (This is not good mathematics, please don’t quote it).

Field Programmable Gate Arrays

Another less publicised beneficiary is the family of silicon devices known as FPGAs. These relatively new devices contain uncommitted silicon resources that may be configured to perform any function the designer wishes. Crucially, they may be reconfigured any number of times which means that more or less anyone can develop their own bespoke silicon chips without the inconvenience of having to build a wafer fabrication plant. In doing so, FPGA vendors have created a viable alternative to the traditional approach to high-performance computing which has been to formulate general purpose processing elements into parallel structures on proprietary data highways – so called cluster programming. But it is difficult and expensive to do this efficiently – and I think it is self-evident that nothing is as efficient as creating a chip specifically for your application – in Calrec’s case, processing audio.

Calrec has made dramatic use of this technology in digital mixing console designs. The Alpha products started by off-loading the repetitive numerical processing tasks to FPGAs in order to save DSP cycles. Later, all DSP processing was transferred to a small number of FPGA chips, realising a massive increase in efficiency. Moore’s Law increases in FPGA capacity have meant that the latest DSP design, Bluefin2, packs all the audio processing for more than 1,000 channels and hundreds of output buses into just four FPGA chips. This reduces power requirements, meaning smaller power supplies and less heat to manage, but, most significantly, the huge reduction in components and connectors means better reliability.

Of course, all this powerful FPGA silicon takes a great deal of programming. There’s no doubt that development tools have improved, but we can’t expect to find a programming language that frees us from the burden of clarifying our ideas. Perhaps this is the downside to FPGA, as designs are most efficiently realised in less abstract languages. We do not benefit from a Moore’s law of software efficiency.

Interestingly, the problem appears to be the reverse for general purpose processors – as chips have become more powerful and storage density increased, there has been a tendency for the installation footprint of newer programs to increase. Computers come with masses of pre-installed bloatware or foistware, mostly there to market trial versions of unwanted programs. Another cause of unwanted software is the proliferation of competing standards all of which require support. Inevitably, the situation has a law to describe it – Wirth’s Law, which states that ‘software speed is decreasing more quickly than hardware speed is increasing’.

The attraction of Moore’s Law is not the uncanny accuracy of the prediction, now 45 years old. Rather, it embodies the self-confidence that is the driving force behind half a century of mind-spinning technological innovation. It is the mantra of an industry in such supreme state of health that it leaps technological hurdles that would stop others in their tracks, time after time, like some industrial Usain Bolt on titanium springs.

The current prediction is that Moore’s Law will last until 2015. My suspicion is that it will continue far beyond. Although the original claim refers only to silicon, and it may be the case that silicon has run its race, there are alternative technologies lined up waiting to pick up the baton. Perhaps optical, chemical or even quantum computing will be the proud defender of Moore’s law for the next half century.

If it does, the possibilities are interesting. Some more dubious maths – if a 1,000-channel broadcast mixing console uses about 3.3bn transistors in its DSP processing, it might comfortably fit into two large currently available FPGA devices. If Moore’s law continues to apply, in just eight years' time, the same chips would provide 256,000 channels of processing – at least enough for all the broadcast consoles in the United Kingdom.

Constant beauty

There is a deep inspiring beauty in the application of mathematics where it might least be expected; the more arcane the better. Seth Lloyd, a quantum mechanic at MIT shows how the potential computing capacity of a kilogram of matter equals pi times energy divided by Planck’s Constant. This is a big number. About 5.0 x 1050 operations per second. This is enough to provide everyone on planet earth with 2,000 times more processing power than there is in the world today. What possible calculation could better this? The FFT of a kiss, the z-plane transform of disappointment?

There is a small group of futurists who believe that Moore’s law style improvements in technology will eventually lead to the acceleration of intelligence, which itself will lead to an intelligence explosion and a technological singularity, beyond which, the state of the world or the human condition cannot be predicted. Perhaps Armageddon, or maybe a matrix-like virtual world where machines rule our minds.

Are these people mad? Undoubtedly. But could they be right? Let’s hope not. Best not to worry about it too much. I know I can’t – I have to go and bake an internet-shaped birthday cake.

Patrick Warrington is Technical Director at Calrec

TwitterGoogle BookmarksRedditLinkedIn Pin It
Fast-and-Wide.com An independent news site and blog for professional audio and related businesses, Fast-and-Wide.com provides a platform for discussion and information exchange in one of the world's fastest-moving technology-based industries.
Fast Touch:
Author: This email address is being protected from spambots. You need JavaScript enabled to view it.

 
Fast Thinking:Marketing:  This email address is being protected from spambots. You need JavaScript enabled to view it.
Web: Latitude Hosting