Stories
Slash Boxes
Comments

SoylentNews is people

posted by Fnord666 on Sunday July 12 2020, @08:32AM   Printer-friendly
from the going-analog dept.

Imec Develops Efficient Processor In Memory Technique for GloFo

Imec and GlobalFoundries have demonstrated a processor-in-memory chip that can achieve energy efficiency up to 2900 TOPS/W, approximately two orders of magnitude above today's commercial processor-in-memory chips. The chip uses an established idea, analog computing, implemented in SRAM in GlobalFoundries' 22nm fully-depleted silicon-on-insulator (FD-SOI) process technology. Imec's analog in-memory compute (AiMC) will be available to GlobalFoundries customers as a feature that can be implemented on the company's 22FDX platform.

Since a neural network model may have tens or hundreds of millions of weights, sending data back and forth between the memory and the processor is inefficient. Analog computing uses a memory array to store the weights and also perform multiply-accumulate (MAC) operations, so there is no memory-to-processor transfer needed. Each memristor element (perhaps a ReRAM cell) has its conductance programmed to an analog level which is proportional to the required weight.

[...] Imec has built a test chip, called analog inference accelerator (AnIA), based on GlobalFoundries' 22nm FD-SOI process. AnIA's 512k array of SRAM cells plus digital infrastructure including 1024 DACs and 512 ADCs takes up 4mm2. It can perform around half a million computations per operation cycle based on 6-bit (plus sign bit) input activations, ternary weights (-1, 0, +1) and 6-bit outputs.

[...] Imec showed accuracy results for object recognition inference on the CIFAR 10 dataset which dropped only one percentage point compared to a similarly quantised baseline. With a supply voltage of 0.8 V, AnIA's energy efficiency is between 1050 and 1500 TOPS/W at 23.5 TOPS. For 0.6 V supply voltage, AnIA achieved 5.8 TOPS at around 1800-2900 TOPS/W.

Promising application: edge computing facial recognition cameras for the surveillance state.

Also at Wccftech.

See also: Week In Review: Auto, Security, Pervasive Computing

Previously: IBM Reduces Neural Network Energy Consumption Using Analog Memory and Non-Von Neumann Architecture

Related: "3nm" Test Chip Taped Out by Imec and Cadence
GlobalFoundries Abandons "7nm LP" Node, TSMC and Samsung to Pick Up the Slack - "The manufacturer will continue to cooperate with IMEC, which works on a broader set of technologies that will be useful for GF's upcoming specialized fabrication processes..."
Radar for Your Wrist


Original Submission

Related Stories

"3nm" Test Chip Taped Out by Imec and Cadence 13 comments

Imec and Cadence Tape Out Industry's First 3nm Test Chip

The world-leading research and innovation hub in nanoelectronics and digital technologies, imec, and Cadence Design Systems, Inc. today announced that its extensive, long-standing collaboration has resulted in the industry's first 3nm test chip tapeout. The tapeout project, geared toward advancing 3nm chip design, was completed using extreme ultraviolet (EUV) and 193 immersion (193i) lithography-oriented design rules and the Cadence® Innovus™ Implementation System and Genus™ Synthesis Solution. Imec utilized a common industry 64-bit CPU for the test chip with a custom 3nm standard cell library and a TRIM metal flow, where the routing pitch was reduced to 21nm. Together, Cadence and imec have enabled the 3nm implementation flow to be fully validated in preparation for next-generation design innovation.

A tape-out is the final step before the design is sent to be fabricated.

Meanwhile, Imec is looking towards nodes smaller than "3nm":

[...] industry is pinpointing and narrowing down the transistor options for the next major nodes after 3nm. Those two nodes, called 2.5nm and 1.5nm, are slated to appear in 2027 and 2030, respectively, according to the International Technology Roadmap for Semiconductors (ITRS) version 2.0. Another organization, Imec, is more aggressive with the timetable, saying that 2.5nm or thereabouts will arrive by 2024.

Also at EE Times.

Related: TSMC Plans New Fab for 3nm
Samsung Plans a "4nm" Process
IBM Demonstrates 5nm Chip With Horizontal Gate-All-Around Transistors
GlobalFoundries to Spend $10-12 Billion on a 7nm Fab, Possibly $14-18 Billion for 5nm
TSMC Holds Groundbreaking Ceremony for "5nm" Fab, Production to Begin in 2020


Original Submission

IBM Reduces Neural Network Energy Consumption Using Analog Memory and Non-Von Neumann Architecture 38 comments

IBM researchers use analog memory to train deep neural networks faster and more efficiently

Deep neural networks normally require fast, powerful graphical processing unit (GPU) hardware accelerators to support the needed high speed and computational accuracy — such as the GPU devices used in the just-announced Summit supercomputer. But GPUs are highly energy-intensive, making their use expensive and limiting their future growth, the researchers explain in a recent paper published in Nature.

Instead, the IBM researchers used large arrays of non-volatile analog memory devices (which use continuously variable signals rather than binary 0s and 1s) to perform computations. Those arrays allowed the researchers to create, in hardware, the same scale and precision of AI calculations that are achieved by more energy-intensive systems in software, but running hundreds of times faster and at hundreds of times lower power — without sacrificing the ability to create deep learning systems.

The trick was to replace conventional von Neumann architecture, which is "constrained by the time and energy spent moving data back and forth between the memory and the processor (the 'von Neumann bottleneck')," the researchers explain in the paper. "By contrast, in a non-von Neumann scheme, computing is done at the location of the data [in memory], with the strengths of the synaptic connections (the 'weights') stored and adjusted directly in memory.

Equivalent-accuracy accelerated neural-network training using analogue memory (DOI: 10.1038/s41586-018-0180-5) (DX)


Original Submission

GlobalFoundries Abandons "7nm LP" Node, TSMC and Samsung to Pick Up the Slack 15 comments

GlobalFoundries has halted development of its "7nm" low power node, will fire 5% of its staff, and will also halt most development of smaller nodes (such as "5nm" and "3nm"):

GlobalFoundries on Monday announced an important strategy shift. The contract maker of semiconductors decided to cease development of bleeding edge manufacturing technologies and stop all work on its 7LP (7 nm) fabrication processes, which will not be used for any client. Instead, the company will focus on specialized process technologies for clients in emerging high-growth markets. These technologies will initially be based on the company's 14LPP/12LP platform and will include RF, embedded memory, and low power features. Because of the strategy shift, GF will cut 5% of its staff as well as renegotiate its WSA and IP-related deals with AMD and IBM. In a bid to understand more what is going on, we sat down with Gary Patton, CTO of GlobalFoundries.

[...] Along with the cancellation of the 7LP, GlobalFoundries essentially canned all pathfinding and research operations for 5 nm and 3 nm nodes. The company will continue to work with the IBM Research Alliance (in Albany, NY) until the end of this year, but GlobalFoundries is not sure it makes sense to invest in R&D for 'bleeding edge' nodes given that it does not plan to use them any time soon. The manufacturer will continue to cooperate with IMEC, which works on a broader set of technologies that will be useful for GF's upcoming specialized fabrication processes, but obviously it will refocus its priorities there as well (more on GF's future process technologies later in this article).

So, the key takeaway here is that while the 7LP platform was a bit behind TSMC's CLN7FF when it comes to HVM – and GlobalFoundries has never been first to market with leading edge bulk manufacturing technologies anyway – there were no issues with the fabrication process itself. Rather there were deeper economic reasons behind the decision.

GlobalFoundries would have needed to use deep ultraviolet (DUV) instead of extreme ultraviolet (EUV) lithography for its initial "7nm" chips. It would have also required billions of dollars of investment to succeed on the "7nm" node, only to make less "7nm" chips than its competitors. The change in plans will require further renegotiation of GlobalFoundries' and AMD's Wafer Supply Agreement (WSA).

Meanwhile, AMD will move most of its business over to TSMC, although it may consider using Samsung:

Radar for Your Wrist 11 comments

IEEE Spectrum:

Tired of trying to tap icons on small smartwatch screens? Some day you could just swipe right through the air above them thanks to miniaturized radar technology and its accompanying gesture recognition technology in development at imec, the Belgium-based R&D center.

Imec's radar chips operate at around 145 GHz, well above the bands used for car radar. That high up in the electromagnetic spectrum, the chip can take advantage of a full 10 gigahertz of bandwidth, which leads to millimeter-scale resolution, its inventors say.

"Gestures allow a lot of capabilities where screens are becoming too small for fine movements," says Kathleen Philips, program director for IoT at imec. "Radar is great for measuring movement; this particular radar is great for measuring micromovements."

If they combine this technology with a wand that must be swept in precise patterns, they will be richer than God.


Original Submission

This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
(1)
  • (Score: 0) by Anonymous Coward on Sunday July 12 2020, @01:18PM (4 children)

    by Anonymous Coward on Sunday July 12 2020, @01:18PM (#1019820)

    Anyone heard of John Gufstaffson?

  • (Score: 2, Interesting) by Anonymous Coward on Sunday July 12 2020, @05:20PM (4 children)

    by Anonymous Coward on Sunday July 12 2020, @05:20PM (#1019908)

    How do these processors-in-memory compare to traditional processor/memory combo in neural network implementation?

    • (Score: 4, Funny) by takyon on Sunday July 12 2020, @06:31PM (3 children)

      by takyon (881) <takyonNO@SPAMsoylentnews.org> on Sunday July 12 2020, @06:31PM (#1019948) Journal

      I'm not sure, but I believe Intel Xeon Phi and Nvidia GPUs are around the 0.5-2 TOPS/W range. An article about the Groq TSP [techspot.com] gives its performance at 1000 TOPS / 300 Watts, vs. 250 TOPS / 300 Watts or 130 TOPS / 70 Watts for two Nvidia products.

      PIM Techniques Boost AI Inference to 8.8 TOPS/W [eetimes.com]

      This is also described as processor-in-memory, but only 8.8 TOPS/W.

      Gyrfalcon’s New Chip Raises Bar (12.6 TOPS/W) on High Performance Edge AI with Lower Power Use [gyrfalcontech.ai]

      Edge AI chip forgoes multiply-accumulate array to reach 55 TOPS/W [embedded.com]

      1050-2900 TOPS/W, or 1-3 exaOPS/Watt, is just incredibly efficient, and they are apparently aiming for 10 exaOPS/Watt in a future version (the article says "10 TOPS below 100 mW" but I think they meant "below 1 mW"). The actual performance of 5.8 or 23.5 TOPS is not bad if it can be used by a single camera, for instance. That 55 TOPS/W chip mentioned above "can run the equivalent of 4 TOPS" in order to do "YOLOv3 at 30fps" "in under 20 mW". YOLOv3 is a real-time object detection algorithm.

      So you can see how power sipping inference technology would be extremely useful for a surveillance state. Just make tiny, cheap cameras that store facial signatures or facial snapshots with timestamps indefinitely, instead of recording a limited number of hours/days of video. Power the camera and AI chip combo with solar panels or ambient energy [networkworld.com]. Deploy millions of them everywhere, even in small towns, download from them as needed. Full video footage will be more useful in some situations, but this could be a cheap way to track the movements of every single person.

      --
      [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
      • (Score: 1, Touché) by Anonymous Coward on Sunday July 12 2020, @06:56PM (2 children)

        by Anonymous Coward on Sunday July 12 2020, @06:56PM (#1019966)

        "Deploy millions of them everywhere, even in small towns"
        Hey, thanks for giving them ideas.

        • (Score: 3, Insightful) by takyon on Sunday July 12 2020, @07:27PM (1 child)

          by takyon (881) <takyonNO@SPAMsoylentnews.org> on Sunday July 12 2020, @07:27PM (#1019979) Journal

          It's just the natural progression of the CCTV craze. Less human effort needed to make it useful, with more real-time AI features. Maybe all participating cameras within a 100 mile search radius can be sent your signature and ping back only if you (or someone else on the shit list) are seen. Cameras in sensitive high-traffic areas will gain features like "emotion recognition" [standard.co.uk].

          Ultra low power capabilities will make it easier to deploy in small towns or rural areas, possibly with no need for access to the power grid. So villages in China, Alaska, etc. highway markers, wherever needed. Nobody will be immune. And they should definitely be attached to all 5G microcells.

          --
          [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
(1)