feature article
Subscribe Now

Altera Looks Forward

Insight from an Industry Leader

In the rough-and-tumble, day-to-day, my-chip’s-bigger-than-your-chip schoolyard scrap that characterizes the top tier of the FPGA industry, a glimpse of vision, long-term insight and strategy are a rare breath of fresh air. We often feel that the two toughest competitors in the business spend too much time staring each other down and not enough time strategizing on how to conquer more of the vast landscape of logic design opportunity waiting patiently at the forefronts of their fiefdoms. However, when we sat down this week with Danny Biran, Altera’s Vice President of Product and Corporate Marketing, vision and insight are exactly what we got.

It seems that Altera has donned some panoramic goggles and is surveying the silicon situation with a calm and rational head, looking at the emerging role of programmable logic in the broader electronics industry. Altera’s vision starts at the very root of our core assumptions about modern electronics systems – the distinction between hardware and software. “We have all been programmed to believe that hardware is difficult to change and software is easy to change,” observes Biran, “but programmable logic breaks those stereotypes.” Of course, we’ve all become accustomed to the assumption that hardware is expensive to develop, has long development times, and is difficult to change once designed. Software is easier and faster to change, simpler to develop, and comparatively easy to deploy. We also generally assume and accept that software can be updated after the product is in the customer’s hands while hardware cannot.

FPGAs change those well-established rules, however. Hardware can be modified rather easily, even in the field. The death-defying challenge of ASIC verification melts away, and the choice of implementing algorithms as processor instructions or logic elements becomes a more complex tradeoff of performance, power consumption, silicon area, and design time dedicated to each module. The line between hardware and software becomes ever so fuzzy. If the concepts of hardware and software are no longer sacrosanct, what exactly does Altera see in the new digital design domain?

Altera sees that systems that formerly were almost completely defined in hardware have grown to include large software components. Operating systems were added to insulate the complexity of the software from changes in the underlying hardware platform and to manage memory and interrupts. Hardware architectures grew more complex as well, and performance expectations increased almost exponentially. At the same time, design cycles and product life cycles decreased dramatically. Systems companies were faced with the challenge of creating and verifying increasingly complex systems on shorter schedules in order to hit ever-diminishing market windows.

At the same time, the cost of deploying the predominant custom silicon technology – cell-based ASIC, is increasing with each new process generation. This has made the ASIC option evolve into the purview of the few, available only to companies with the resources, experience, and product volume to justify the astronomical development costs and risks. The rest of the industry was left to come up with other options like combinations of ASSPs, FPGAs, and functions that could be implemented in embedded software running on standard processing platforms. Over time, FPGAs have grown in capacity, performance, and capability, subsuming more and more of the functionality of the system and providing software-like flexibility to the portion of the system that they encompassed.

Designers soon began to view FPGAs as something more than a simple substitute for ASIC in lower-volume or lower-budget applications. By leveraging reprogrammability, teams are designing products that get to market faster and stay in the field longer. Altera sees field programmability as migrating from a “nice to have” to a “must” for companies wanting return on their system development investments. “I don’t know of a market that isn’t going through this,” continues Biran. “No one has the luxury anymore in any established market to wring out the profit from a hardware design for five years. No sooner is a design frozen than cost reduction and feature enhancement programs are underway.”

At the same time, Altera observes that globalization hasn’t conquered the standards dilemma. If you’re introducing a new product for the global market, you’ll find that what seems like it should be a single silicon implementation may multiply into a mess of variants for different geographic regions. Even if your overall production volume might justify an ASIC, by the time you segment into a different version for each different networking standard and for each geography, you may have created ten mid-volume products from a single high-volume one. With programmable logic, however, we have the opportunity to create a single hardware platform that can be configured differently for each variant. Ironically, by using what is normally considered a lower-volume technology (FPGA) we can push the hardware portion of our design back into the high-volume arena.

This product fragmentation goes far beyond simple geographic lines, however. One of the hottest new markets for FPGAs is computer and television displays. These products are offered in a daunting array of sizes and configurations, taking into account screen size and resolution, performance, HDTV and broadcast standards, and geographic target. FPGAs can bring an almost unmanageable problem back into the realm of sanity. Without programmable logic, many product variations simply could not exist due to the economics of creating custom silicon for lower-demand models.

The final points of Biran’s programmable logic prognostication relate to the topic of performance. Altera observes that processors have hit a wall in their performance progression, and that processor companies are rapidly migrating toward dual-core and multi-core solutions to circumnavigate the barrier. The same trend is emerging whether we look at conventional general-purpose processor architectures or more domain-specific processors such as DSPs. There is widespread industry agreement that parallelism and not increased clock frequency will account for the next several rounds of performance improvement in computing.

If that is true, Altera argues, the processing platform of the future will include billions of transistors on a single chip, reconfigurable circuit blocks, optimized special-purpose blocks for real-time signal processing, on-chip memory, high-speed interconnect, and compatibility with existing software. Altera points out that this hardware configuration is exactly what we already have in leading FPGAs today. Altera makes a good case. While Moore’s Law may take us up and to the right in both density and performance, the basic modern FPGA hardware architecture is a logical evolution of the track that conventional processors, DSPs, and programmable logic have been following for the past few years.

Altera’s conclusion matches that of many other groups with whom we’ve consulted about the future of computing. All roads ultimately lead to a reconfigurable hardware platform with the hardware/software partitioning decision based on performance, power, and real-estate factors. If true, we interpret this trend to mean that the key enabling technology in electronics for the next couple of decades will be design tools, from conventional compilers and electronic design automation aids to next-generation high-level language synthesis and hardware/software partitioning technology.

Today, the biggest single barrier to widespread adoption of FPGAs as algorithm accelerators is the immaturity of the design methodology. The average system designer just doesn’t want to incur the time and expense of producing a register-transfer level VHDL or Verilog design in order to get key components of their system implemented in FPGA fabric. EDA and FPGA vendors are chipping away at that problem with IP-based design methodologies, C/C++ synthesis, Matlab/Simulink integration, and a number of novel and innovative approaches that will make the power of the FPGA fabric available to a wider audience. Ultimately, though, these tools and methods still have a long way to go.

Both tools and engineers must evolve to meet the challenges of the coming decade. As the barriers between hardware and software blur and ultimately disappear, so must the distinction between hardware and software engineers. Lest any of us get too comfortable in our roles, we should be mindful of the history-proven fact that the only constant in our industry is change. That change knows no barriers and observes no established definitions of technology, tools, or job functions. As long as we understand that engineers’ true expertise must always be creating and adapting to that change, we’ll survive and flourish. Evidently, Altera agrees.

Leave a Reply

featured blogs
Nov 22, 2024
We're providing every session and keynote from Works With 2024 on-demand. It's the only place wireless IoT developers can access hands-on training for free....
Nov 22, 2024
I just saw a video on YouTube'”it's a few very funny minutes from a show by an engineer who transitioned into being a comedian...

featured video

Introducing FPGAi – Innovations Unlocked by AI-enabled FPGAs

Sponsored by Intel

Altera Innovators Day presentation by Ilya Ganusov showing the advantages of FPGAs for implementing AI-based Systems. See additional videos on AI and other Altera Innovators Day in Altera’s YouTube channel playlists.

Learn more about FPGAs for Artificial Intelligence here

featured paper

Quantized Neural Networks for FPGA Inference

Sponsored by Intel

Implementing a low precision network in FPGA hardware for efficient inferencing provides numerous advantages when it comes to meeting demanding specifications. The increased flexibility allows optimization of throughput, overall power consumption, resource usage, device size, TOPs/watt, and deterministic latency. These are important benefits where scaling and efficiency are inherent requirements of the application.

Click to read more

featured chalk talk

Versatile S32G3 Processors for Automotive and Beyond
In this episode of Chalk Talk, Amelia Dalton and Brian Carlson from NXP investigate NXP’s S32G3 vehicle network processors that combine ASIL D safety, hardware security, high-performance real-time and application processing and network acceleration. They explore how these processors support many vehicle needs simultaneously, the specific benefits they bring to autonomous drive and ADAS applications, and how you can get started developing with these processors today.
Jul 24, 2024
91,805 views