News

January 18, 2024

MicroTCA.0 Revision 3 Delivers 4x Performance Improvement with 100 Gigabit Ethernet, PCIe Gen 5 in up to 12 Slots

Industry NewsMicroTCAMicroTCANewsPICMG

Latest MicroTCA revision adds high-bandwidth interfaces and increased TDP to prepare the specification for demanding next-generation edge and server workloads

WAKEFIELD, MA. PICMG, a leading consortium for the development of open embedded computing specifications, has ratified Revision 3.0 of the MicroTCA.0 (µTCA.0) specification. This new release addresses urgent bandwidth requirements by defining 100 GbE and PCIe Gen 5 fabrics that improve system performance by 4x while also expanding platform thermal design power (TDP), enabling the use of higher performance processors. Users can now select from a range of specification-compliant building blocks and solutions that allow for more power per slot for higher transfer rates within µTCA chassis and to external systems and server clusters. It also lays the foundation for building next-generation MicroTCA proofs of concept.

Revision 3 of MTCA.0 addresses the current and future demands of applications such as machine vision, AI, defense, research, instrumentation, wireless communication, and emerging applications like quantum computing. “Revision 3 of MTCA.0 addresses urgent requirements and thus enables companies to now provide specification-compliant solutions rather than proprietary or custom approaches,” says Heiko Korte of NAT Europe and lead of PICMG’s MicroTCA Technical Working Group. “The fact that so many MicroTCA ecosystem suppliers have joined the working group shows the importance of the changes and also underlines the strong interest to make these part of an open specification.

“The broad spectrum of participants also ensured that every single agenda item got reviewed from different angles and properly discussed,” he adds.

MicroTCA.0 Revision 3.0-compliant solutions will be available shortly, including chassis, MicroTCA Carrier Hub (MCH), Advanced Mezzanine Card (AMC), Rear Transition Module (RTM), and power modules (PM) products from VadaTech, NAT, AIES Sp z o.o., nVent SCHROFF, and others. The commitment of existing vendors to continuing the development of µTCA products is joined by increased interest from players in emerging fields like quantum computing, projecting a healthy lifecycle for the open computing specification for years to come.

“VadaTech is proud to have been part of the development and ratification of revision 3 of the uTCA specification,” says Alex Malcom, Managing Director of VadaTech Ltd. “Its release secures the continued adoption of the standard by commercial, scientific and defense organizations around the world.”

The MicroTCA.0 Revision 3.0 specification can be accessed at www.picmg.org/product/micro-telecommunications-computing-architecture-base-specification. For more information on the MicroTCA family, visit https://www.picmg.org/openstandards/microtca.

About PICMG

Founded in 1994, PICMG is a not-for-profit 501(c) consortium of companies and organizations that collaboratively develop open standards for high performance industrial, Industrial IoT, military & aerospace, telecommunications, test & measurement, medical, and general-purpose embedded computing applications. There are over 130 member companies that specialize in a wide range of technical disciplines, including mechanical and thermal design, single board computer design, high-speed signaling design and analysis, networking expertise, backplane, and packaging design, power management, high availability software and comprehensive system management.

 

Key standards families developed by PICMG include COM-HPC, COM Express, CompactPCI, AdvancedTCA, MicroTCA, AdvancedMC, CompactPCI Serial, SHB Express, MicroSAM, and HPM (Hardware Platform Management). https://www.picmg.org.

December 20, 2023

congatec launches COM Express Compact module with brand new Intel® Core™ Ultra processors

COM-HPCCongatecIndustry NewsMember NewsNews

Next-generation AI computing for the edge 

Deggendorf, Germany, 20 December 2023 * * * congatec – a leading vendor of embedded and edge computing technology – is introducing its latest range of COM Express Compact modules based on the Intel® Core™ Ultra processors. Providing a unique combination of heterogeneous compute engines, including CPU, GPU and NPU, the new modules are an ideal fit to run demanding AI workloads at the edge.

Next to the powerful P-cores and efficient E-cores for general computing and the high-performance Intel® Arc™ GPU for graphics-intensive tasks, the integrated Neural Processing Unit (NPU) called Intel® AI Boost contributes advanced neural processing capabilities to the overall computational architecture. The integrated NPU enables highly efficient integration of advanced artificial intelligence workloads at lower system complexity and costs than discrete accelerators. This makes the new Intel Core Ultra processor-based Computer-on-Modules especially beneficial for combining high-performance real-time computing with powerful AI capabilities in surgery robots and medical imaging and diagnostic systems, where automatically generated critical findings can support medical personnel. Other application targets are situational awareness in industrial applications such as inspection systems, stationary robotic arms, autonomous mobile robots (AMRs), and autonomous guided vehicles (AGVs), to name just a few.

“The new conga-TC700 COM Express Compact Computer-on-Modules provide application-ready AI capabilities in a plug-and-play form factor. Its ecosystem of customer solution-focused products and services significantly improves the time-to-market for implementing the latest state-of-the-art x86 with powerful AI capabilities as required in industrial process control, microgrid controller, medical ultrasound and x-ray, automated check-out terminals, powerful AMRs, and many more,” says Tim Henrichs, VP Global Marketing & Business Development at congatec. “OEMs can simply upgrade existing applications by exchanging the module and instantly gain access to cutting-edge AI technologies. It has never been easier to integrate artificial intelligence in x86-based systems.”

 

The feature set at a glance

The new conga-TC700 COM Express Compact Computer-on-Modules with Intel Core Ultra processors (code named Meteor Lake) are among the most power-efficient x86 client SoCs available on the market. Up to 6 P-Cores, up to 8 E-Cores, and 2 Low Power E-Cores support up to 22 threads, making it possible to consolidate distributed devices onto a single platform for the lowest total cost of ownership. The SoC-integrated Intel Arc GPU with up to 8 Xe Cores and up to 128 EUs can handle stunning graphics up to 2x 8K resolution and ultra-fast GPGPU-based vision data (pre)processing. The integrated NPU Intel AI Boost executes machine learning algorithms and AI inferences particularly efficient. Up to 96 GB DDR SO-DIMM with in-band ECC at 5600 MT/s contributes to power-efficient high data throughput and low latency.

The modules are supported by congatec’s OEM solution-focused high-performance ecosystem, including highly efficient active and passive cooling solutions and ready-to-use evaluation carrier boards. Customers can order the modules with pre-evaluated real-time hypervisor technology from Real-Time Systems for virtual machine deployments and workload consolidation in edge computing scenarios. Service offerings include shock and vibration tests for custom system designs, temperature screening, and high-speed signal compliance testing, along with design-in services and all required training sessions to simplify the use of congatec’s embedded computer technologies to round off the ecosystem.

The new conga-TC700 COM Express Compact Type 6 modules support the embedded temperature range from 0 °C to 60 °C and are available in the following standard configurations:

Processor P-cores /
E-cores / Threads
Max. Turbo Freq. [GHz]
P-cores /
E-cores
Base Freq. [GHz]
P-cores /
E-cores
Intel Smart Cache [MB] Graphics [Execution Units] CPU Base Power [W]
             
Intel Core Ultra 7 155H processor 6/10/22 4.8 / 3.8 1.4 / 0.9 24 128 28
Intel Core Ultra 7 155U processor 2/10/14 4.8 / 3.8 1.7 / 1.2 12 64 15
Intel Core Ultra 5 125H processor 4/10/18 4.5 / 3.6 1.7 / 1.2 18 112 28
Intel Core Ultra 5 125U processor 2/10/14 4.3 / 3.6 1.3 / 0.8 12 64 15

 

All features of the new Intel Core Ultra processor-based conga-TC700 COM Express Compact modules can be found here: https://www.congatec.com/en/products/com-express-type-6/conga-tc700/

For more information on the innovative Intel Core Ultra processor platform please visit https://www.congatec.com/en/technologies/intel-meteor-lake-h-based-computer-on-modules/

You can experience these and other innovations at embedded world from 9-11 April 2024: https://www.congatec.com/en/congatec/events/congatec-at-embedded-world-2024/

Visit congatec in Hall 3 at Stand 241

* * *

About congatec 

congatec is a rapidly growing technology company focusing on embedded and edge computing products and services. The high-performance computer modules are used in a wide range of applications and devices in industrial automation, medical technology, robotics, telecommunications and many other verticals. Backed by controlling shareholder DBAG Fund VIII, a German midmarket fund focusing on growing industrial businesses, congatec has the financing and M&A experience to take advantage of these expanding market opportunities. congatec is the global market leader in the computer-on-modules segment with an excellent customer base from start-ups to international blue chip companies. More information is available on our website at www.congatec.com or via LinkedIn, X (Twitter) and YouTube.

November 29, 2023

Cloud-Native Gets An Edge with COM-HPC Server Modules and Manycore Arm SoCs

ADLINKCOM-HPCOpen StandardsSlider

PICMG executive member ADLINK Technology is redefining the performance per watt equation in computer-on-module technology with COM-HPC Server Type modules based on Ampere Computing’s “cloud-native” manycore Arm processors. ADLINK Technology’s business development manager for embedded modules, Richard Pinnow, and Joe Speed, Head of Edge at Ampere Computing, explain.

PICMG: What is a “cloud-native processor” and why does industry need it?

SPEED, AMPERE: Ampere’s founders are the people who created processors for the cloud business at Intel. When you look at cloud, the basic exercise is I take each core of a CPU and sell it to a different customer. This is a gross oversimplification, but you get the idea.

To do this you need privacy – you need to know that one customer can’t see into another customer’s business. You need what’s called “freedom from interference,” so customers can run their own workloads with predictably fast response times regardless of what activity others are doing on the same processor.

Everything is virtualized, so each core is virtualized running in its own OS instance or, if they bought a bunch of cores, their application is containerized running different parts of their application on different cores. You need to be able to do this in an energy-efficient way because it’s not just about having more cores-per-socket, it’s about having more cores-per-rack-per-data center using less power.

I started working with Ampere years ago when I was at ADLINK and for me it was all about robotics, industrial, software-defined vehicles, and autonomous driving. So, I can get a predictably good response time, right? For me that’s like deterministic, real-time, low jitter, right? Freedom from interference? I can have my autonomous driving perception pipeline feeding sensor fusion data to localization, path planning, and control algorithms and I can pin all of those to separate cores and even run different kinds of software on different bundles of cores and have this very deterministic real time behavior.

So for me taking cloud-native to the edge just belongs.

PICMG: What’s driving demand for cloud-native technology?

PINNOW, ADLINK: It’s definitely scalability, cost efficiency, and time-to-market or agility.

Customers often require, from the application point of view, the ability to scale, and these services need to be very adaptive. Cloud-native processing allows this scalability, especially when we’re talking about the emphasis of microservices on containers, which allows for very dynamic scaling no matter the underlying hardware. And that’s not just focused on the traditional server market – it applies to the edge devices as well.

In terms of cost efficiency, cloud-native technologies enable customers to pay only when resources are allocated, which can lead to significant cost savings. The flexibility to allocate and deallocate resources on demand. And the third pillar is time-to-market, so it’s very important to rapidly develop PoCs and so on, and cloud-native processing really makes the best use of practices like continuous integration and continuous deployment concepts and pipelines. This is getting more and more important for edge devices as well, as it enables faster development cycles and quicker time to market. So the embedded market is adapting the concepts driven by cloud-native processing.

SPEED, AMPERE: There are stats that say 75% of all data is created at the edge, so you have to move the compute to where the data is. If 75% of data is being generated at the edge, you can sure eat a lot of cores pretty quickly, right? And the thing that was really eye opening for a lot of the cloud providers was computer vision because that one is especially greedy in terms of compute resources and communication bandwidth. Backhauling video to the cloud for processing is an expensive proposition. And for a lot of these use cases, by the time it hits the cloud, runs your analysis, and the decision comes back, it’s too late.

And a lot of these use cases talk about “connected” devices, but as you know, we should think of these as “usually connected” or “occasionally connected” devices. If it’s doing a safety function then it has to always work whether the Internet connection is working or not.

We’re Arm-based and we believe in the Arm architecture. A lot of the success of Arm comes from mobile devices that are power constrained by definition. The performance of most Arm products kind of falls off where we begin, and then you get into overlapping with x86, which is hot and hungry compared to what we do on embedded and the edge. With the Ampere Altra we’ve done some recent benchmarks running Yolo V8 AI streams on our chip and it does 4x as many frames per second per watt compared to a top-of-the-line Xeon D processor, so you can support four times as many cameras on the same power budget.

It’s an architecture that’s extremely energy-efficient and then you can scale up to where you’ve got 32, 64, 96, 128, or even 192 cores with our newest product, with each core being very energy efficient and having freedom from interference. They are single-threaded cores and they’re running at a fixed frequency so there’s none of this frequency scaling. That’s the deterministic part so that you get this predictable, low-latency performance. And then as you start loading more cores, everything runs at a fixed, predictable speed.

PICMG: What embedded edge application would make use of 128 cores?

PINNOW, ADLINK: Embedded applications that require 128 cores would typically involve tasks that demand extremely high levels of parallel compute. Some examples are edge devices such as advanced robotics, autonomous vehicles, and industrial automation systems that really benefit from, for instance, real-time computer vision, natural language processing, machine learning, and so on.

To analyze and react to very complex environments rapidly and do all those tasks in parallel

For instance, an autonomous drone might need to process multiple video streams to perform object detection on incoming camera data, simultaneously manage flight controls, and make navigation decisions, all in real time. 128 cores enables you to assign cores to very specific tasks and do all this in parallel without the “bad neighbor effect” where one core or schedule is impacting the calculation of another task or application. A lot of people are using discrete GPU solutions on the market because it allows you to do a lot of tasks in parallel. Here, you basically have the same with 128 cores of unparalleled resolution, and you can select which core is doing what, when. It’s great.

SPEED, AMPERE: When I was the Field CTO at ADLINK we helped launch this thing called “Project Cassini.” Project Cassini is about cloud server virtualization for embedded systems. My friend Garish Shirasat, who was leading the software-defined vehicle efforts over at Arm, had an idea, “What if we put that in a car? It would be Cassini on wheels.” So we used Ampere’s processor to build the developer platform for this software-defined vehicle platform.

Autonomous driving is an obvious workload for that, but what’s happening is all these automakers are working on future silicon they’ll get in a few years but that doesn’t help them develop today. So we took all this work around Project Cassini and worked out how to make it functionally safe and how to put it into cars. What happened is we developed what’s now the Ampere Altra Developer Platform – a 32- to 128-core Arm workstation for developers of Scalable Open Architecture for Embedded Edge (SOAFEE) – which is a big software-defined vehicle program for automakers and automotive tech companies. The Ampere Altra Developer Platform by ADLINK is the reference platform for that.

Figure 1. ADLINK performed eight weeks of testing on the Ampere Altra COM-HPC developer kit, including thermal shock and vibration to MIL-SPEC and validation out to +85 ºC and -45 ºC using a fanless heatsink.

Alternatively, take application code that’s been written for Raspberry Pi and things of that ilk. Jeff Geerling and Patrick Kennedy of Serve the Home did a benchmarking of this telco AI edge server with an Ampere processor from our friends at Super Micro and benchmarked it against a Raspberry Pi cluster. One of our chips was equal to 100 Raspberry Pi 4s in performance, but the interesting thing is this system with redundant 800 W power supplies was still 22% more energy efficient than the Raspberry Pi 4. We have companies working with us where they need to move so many GB per W at 1 W per core. It’s kind of a brilliant fit for those things.

PICMG: The Ampere Altra is a COM-HPC Server Type module and currently the only ADLINK product that supports a “cloud-native” processor, correct? Why that particular PICMG specification and form factor?

PINNOW, ADLINK: Yes, at the moment, the Ampere offering is solely available on COM-HPC and it’s a perfect match for edge computing systems that prioritize energy efficiency and require a lot of scalability. It is outperforming other platforms easily by three times less energy consumption. And from the I/O point of view, the CPU is providing a lot of PCI Express interfaces, way higher memory capacities, and higher bandwidth interfaces so that there are no constraints to get all those signals down to the carrier without a significant loss in signal integrity.

But there’s still a lot of flexibility to interchange from one Ampere Altra COM-HPC module to another or even to an x86 in case it is needed. And not just the hardware is standardized, but firmware is as well. I’m talking, for instance, about the IPMI interface you use to remotely manage COM-HPC devices, regardless if you’re talking to Arm silicon or x86 silicon. The demand driving this market is that application software is getting more and more independent from the underlying architecture.

COM-HPC Ampere Altra | COM-HPC Server Type | COM | ADLINK

Figure 2. The ADLINK Technology, Inc. Ampere Altra COM-HPC Server Type module equips 64 PCIe Gen 4 lanes and six DDR4 channels.

PICMG: COM-HPC was explicitly intended not to not be x86-centric. But we haven’t seen many non-x86-based COM-HPC modules make it to market yet. Why do you think that is?

PINNOW, ADLINK: x86 solutions are and will be complemented by more and more Arm offerings. This is market driven. We cannot avoid this. And the Ampere Altra is a good example.

Our entry-level and super-high-end COMs are Arm based already, Today, right? Now we’re seeing this Ampere Altra being the most powerful COM solution we have at the moment. A very good question is how fast it will cover the traditionally x86-dominated mid-performance market. It depends how fast and reliably a customer can port their existing code to execute regardless of the underlying hardware, which is a key trend. But containers hypervisors and flexible application frameworks are enablers to support this journey. And we see this happening already today. So I think it’s just a matter of time until we see more and more Arm flavors in the mid performance as well.

For more information visit www.ipi.wiki/products/com-hpc-ampere-altra to find some of ADLINK’s Ampere-based products as well as carrier reference designs, schematics, Ethernet OCP cards and design files, and even bill of materials.

Richard Pinnow is business development manager for embedded modules at ADLINK.

Joe Speed is head of Edge at Ampere.

ADLINK Technology, Inc.
www.adlinktech.com/en

Ampere Computing
https://amperecomputing.com