editor's blog
Subscribe Now

Parallel Accurate SPICE

SPICE has got to be one of the oldest tools still being used by designers. So you might expect it to be a mature market, with a few well-established tools battling for the best performance/capacity and/or accuracy (and occasionally even collaborating).

In fact, it’s typically been more about “or” than “and,” as there are generally two SPICE camps: the fast, high-capacity versions that are “good enough” for everyday repeated use as you explore design options, and sign-off-quality versions that are more accurate, but take longer to complete and can’t handle as large a design.

The tradeoffs between the fast/big and accurate versions are usually about simplifying assumptions and models and such. Parallel execution has also helped, although it’s entirely possible that long-in-the-tooth engines were not designed for effective parallelization.

So ProPlus has announced a new SPICE tool, NanoSpice, that leverages its BSIMProPlus high-accuracy engine for analysis of large designs with quick turnaround. They claim they can handle designs of 50-100 million elements 10-100 times faster than competing “traditional” approaches (many of which can’t complete the larger designs in ProPlus’s benchmarking suite).

While they have made some improvements to the performance of the underlying engine, they give most of the credit to parallelization, which scales relatively well (depending on the design – 24 cores giving 8-12x speed-up on most of their examples, with a multiplier design actually achieving around 20x). But what they underscore with this is that it still uses the model that BSIMProPlus uses, suggesting equivalent accuracy.

They also say that they’ve got a better licensing model for using parallelism. Traditional schemes simply use more licenses as you use more machines, but they say that this was largely configured for occasional bursty usage. If everyone is always using parallelism, then you typically run out of licenses that way.

Their solution? Well, I actually don’t know. They are keeping mum about that. So they say it’s different and better; you’ll have to be the judges of that.

You can find out more in their release.

Leave a Reply

featured blogs
Nov 22, 2024
We're providing every session and keynote from Works With 2024 on-demand. It's the only place wireless IoT developers can access hands-on training for free....
Nov 22, 2024
I just saw a video on YouTube'”it's a few very funny minutes from a show by an engineer who transitioned into being a comedian...

featured video

Introducing FPGAi – Innovations Unlocked by AI-enabled FPGAs

Sponsored by Intel

Altera Innovators Day presentation by Ilya Ganusov showing the advantages of FPGAs for implementing AI-based Systems. See additional videos on AI and other Altera Innovators Day in Altera’s YouTube channel playlists.

Learn more about FPGAs for Artificial Intelligence here

featured paper

Quantized Neural Networks for FPGA Inference

Sponsored by Intel

Implementing a low precision network in FPGA hardware for efficient inferencing provides numerous advantages when it comes to meeting demanding specifications. The increased flexibility allows optimization of throughput, overall power consumption, resource usage, device size, TOPs/watt, and deterministic latency. These are important benefits where scaling and efficiency are inherent requirements of the application.

Click to read more

featured chalk talk

Datalogging in Automotive
Sponsored by Infineon
In this episode of Chalk Talk, Amelia Dalton and Harsha Medu from Infineon examine the value of data logging in automotive applications. They also explore the benefits of event data recorders and how these technologies will shape the future of automotive travel.
Jan 2, 2024
58,428 views