S2C.
S2C.

Breaking the SoC lab walls | SemiWiki.com

Breaking the SoC lab walls | SemiWiki.com May 11, 2015
Table of Content [Hide]

    Don Dingee  Published on 05-11-2015 05:00 AM


    There used to be this thing called the "computer lab" with glowing rows of terminals connected to a mainframe or minicomputer. Computers required a lot of care and feeding, with massive cooling and power requirements. Microprocessors and personal computers appeared in the 1970s, with much smaller and less expensive machines placing isolated dots of compute capability on desks of the fortunate ones. When one says computer lab now, they are usually referring to the 100-level computing course at the local community college.


    During the rise of the engineering workstation in the 1980s and 1990s, the mantra was " Sun: the network is the computer" slogan. Ethernet provided an easy way to connect computers quickly, and distributed file systems such as NFS and interprocess communication via sockets or RPCs made distributed application development possible. Compute power scaled easily by adding workstations on the corporate network, and they could be placed anywhere cabling could be run.


    Embedded systems took full advantage, becoming Ethernet-enabled for distributed applications. In the first dot-com and open source boom, vendors with big, expensive hardware based on VME, CompactPCI, AdvancedTCA, and other system standards were creating "virtual labs". These were controlled environments of a hardware and software configuration in a padded cell on the corporate network at headquarters, designed for customers anywhere to log in remotely so they could benchmark code and evaluate the platform.


    SoCs emerged with integrated connectivity, powering a wave of mobile devices and shrinking many embedded platforms to tiny boards. Today, Wi-Fi brings Ethernet everywhere, tablets and laptops bring computers anywhere, and software developers bring code from all over the globe. SoCs now feature multicore processing and dedicated acceleration units and bottleneck-free interconnect to a wider variety of peripherals.


    With all these advances, why is FPGA-based prototyping still stuck on a relatively big, expensive machine in a "SoC lab" with limited access?


    Granted, one cutting-edge large FPGA with high-speed SERDES interfaces is not cheap. Placing several of those large FPGAs in a prototyping system with the proper pin multiplexing and the right partitioning software to chop up SoC designs effectively is an artform, practiced by few. FPGA-based prototyping systems also bring enhanced debugging, many with deep trace features.


    Until now, people using FPGA-based prototyping systems have gotten by with the SoC lab approach. Teams of SoC designers were relatively small, located close enough to walk down to the lab to engage with their masterpiece. IP blocks might be developed and debugged first, then passed into the integrated system for a concentrated, full-up effort.


    All that is changing as SoC complexity continues to increase, third party IP from a variety of sources becomes more prevalent, and software co-verification with advanced operating systems requires expertise from a community scattered across the globe. It is no longer enough to scale the FPGA-based prototyping hardware; the entire approach has to change.


    S2C has outlined the look of a state-of-the-art FPGA-based prototyping platform, and a vision for adding cloud capability, in a new white paper. There is the usual discussion of performance, extensibility, platform-aware synthesis that can partition and pin-multiplex, and debug. Data illustrating a task-level breakdown for today’s SoC design – upwards of $300M by some estimates for large designs on advanced 16/14nm nodes —— shows software and system-level co-verification to be the largest efforts.


    A prototyping system with cloud-based access to management features can transform those efforts. Mixed-level prototyping is now common, with some blocks complete, some under development in RTL, and some in progress using behavioral models. Physically distributed teams are also common, some working on SoC IP, some on operating systems or drivers.


    Cloud-enabled FPGA-based prototyping platforms raise an exciting new set of possibilities for breaking the SoC lab walls. Perhaps a corporate SoC design team coordinates debug efforts with an off-site third-party IP supplier, or brings together software developers from multiple locations to view debug results as they unfold during a verification run. With complexity going up and co-location going down, the approach S2C is suggesting has a great deal of merit.


    Download the entire white paper, along with other S2C white papers, under this title:


    FPGA Prototyping of System-on-Chip Designs

    Back to list Back to list
    Related S2C Complete Prototyping Solutions
    Prodigy S7 Series (Virtex UltraScale+)
    The 7th generation SoC/ASIC prototyping solution from S2C, the Prodigy S7 series Logic System, is equipped with AMD's(Xilinx)Virtex UltraScale+™ FPGA. The Prodigy Logic System are supported by S2C'...
    High Speed MCIO Peripherals
    8-Lane PCIe Root Complex MCIO ModuleProvides one x16 PCIe Slot, only support 8 lanesProvides one 100MHz reference clockProvides 3 LEDs and 1 Push ButtonsOccupies two MCIO connectorsMCIO to PGT Convert...
    Virtex UltraScale Series
    S2C's Virtex UltraScale (VU) Prodigy Logic Systems are built on the Virtex UltraScale XCVU440 FPGA from Xilinx.
    What's New at S2C
    References
    Request for Quote
    What type of chip are you designing
    What is the capacity of the ASIC gate included in the design?
    5 million-20 million
    20 million-50 million
    50 million-100 million
    100 million-1 billion
    More than 1 billion
    Which FPGA do you prefer to use?
    Xilinx VU440
    Xilinx KU115
    Xilinx VU19P
    Xilinx VU13P
    Xilinx VU9P
    AMD VP1802
    AMD VP1902
    Intel S10-10M
    Intel S10-2800
    Not sure, need professional advice
    What kind of FPGA configuration do you need?
    Single FPGA
    Dual FPGA
    Four FPGAs
    Eight FPGAs
    Not sure, need professional advice
    What kind of peripheral interface do you need?
    How many prototype verification platforms do you need?
    Do you need the following tools?
    Segmentation tool
    Multiple FPGA debugging tools
    Co-modeling tool (allows large amounts of data to interact between FPGA and PC host)
    When do you need to use our products?
    0-6 months
    6-12 months
    More than 12 months
    Not sure
    Any additional comments?