Sunday, December 3, 2023
HomeElectronicsThe 4th IBM IEEE CAS/EDS AI Compute Symposium

The 4th IBM IEEE CAS/EDS AI Compute Symposium

//php echo do_shortcode(‘[responsivevoice_button voice=”US English Male” buttontext=”Listen to Post”]’) ?>

The 4th IBM IEEE CAS/EDS AI Compute Symposium was held over two days in October and coated a bunch of subjects starting from cell synthetic intelligence (AI) accelerators, AI–pushed autonomous labs, AI and the cloud, and way more.

The digital symposium, supported by the IBM Academy of Expertise, drew greater than 2400 viewers over two days, participation from 50 nations, over 54 pupil posters, and 11 audio system from business and academia. The theme of the symposium was “From Floor as much as Cloud”. In brief, the symposium coated a spread of subjects from gadget know-how to circuits, structure, and algorithms that pave the best way for cloud improvements, with an emphasis on inexperienced AI.

Keynote speaker Hoi–Jun Yoo, professor of the College of Electrical Engineering and the director of the System Design on the Korea Superior Institute of Science and Expertise, opened the symposium with a presentation associated to “Coaching on Chip — Subsequent Wave of Cellular AI Accelerators”.

Most cell deep neural community (DNN) accelerators goal solely inference of DNN fashions on edge gadgets, whereas on–gadget coaching was out of attain in cell platforms attributable to its extreme computational necessities. Coaching–on–chip (ToC) with person–particular knowledge is changing into extra vital than ever due to privateness points and communication latency of coaching on distant servers.

Yoo highlighted a number of approaches in realizing ToC. Common function {hardware} and software program co–optimization methods aiming to maximise throughput and power–effectivity of DNN coaching had been introduced out with examples, equivalent to sparsity exploitation and bit–precision optimization for coaching. As well as, software particular coaching accelerators for deep reinforcement studying and generative adversarial networks had been mentioned, pertaining to points relating to system implementation with fabricated silicon.

AI–pushed autonomous labs

Teo Laino adopted up with a really fascinating speak about “A Cloud–based mostly AI–pushed Autonomous Lab”. One of the important outcomes of chemistry is the design and manufacturing of latest molecules. The applying of area information accrued over a long time of laboratory expertise has been vital within the synthesis of many new molecular buildings.

Nonetheless, most artificial success tales are accompanied by lengthy hours of repetitive synthesis. Automation techniques had been created lower than 20 years in the past to help chemists with repetitive laboratory duties. Whereas this has confirmed to be very efficient in just a few areas, equivalent to excessive–throughput chemistry, using automation for normal–function duties stays an incredible problem even at this time.

Automation necessitates that chemistry operators write totally different software program for various duties, every of which codifies a particular and distinct sort of chemistry. In the meantime, in natural chemistry, AI has emerged as a helpful complement to human information for duties equivalent to predicting chemical reactions, retrosynthetic routes, and digitizing chemical literature.

Laino’s discuss highlighted the primary cloud–based mostly AI–pushed autonomous laboratory implementation. The AI assists distant chemists with a wide range of duties, together with designing retrosynthetic timber and recommending the proper sequence of operational actions (response circumstances and procedures) or ingesting artificial procedures from literature and changing them into an executable program.

The AI then self–packages the automation layer and makes selections on synthesis execution utilizing suggestions loops from analytical chemistry devices, with supervision from artificial chemists. He introduced the AI core know-how and the way it performs throughout various kinds of artificial duties.

AI and the cloud

Subsequently, Gunnar Hellekson, vice chairman at Purple Hat, spoke about an open method to AI and its integration into the cloud. Enterprise innovation is pushed by huge concepts, transferring quicker than ever earlier than. In the present day, we will do issues we may solely dream of some years in the past. Large world adjustments are shifting the best way folks reside and work, which requires organizations to rethink their groups, processes, and applied sciences to remain aggressive.

In the present day, organizations throughout all geographies and industries can innovate, create extra buyer worth, and differentiation and compete on an equal enjoying discipline. This new actuality calls for that enterprises embrace digital transformation and pivot shortly or fail. AI is a vital a part of the digital transformation journey for a lot of organizations. Sensible cities, wearable well being applied sciences, good power grids, autonomous automobiles, manufacturing, and agriculture are simply a few of the key markets being reworked by AI.

Hellekson identified how know-how, open–supply communities, and new methods of collaborating are driving enterprise improvements like AI. AI investments throughout each business are accelerating to develop differentiated providers and acquire aggressive benefits.

The distinction between advanced {hardware} and software program for key parts like safety and knowledge is diminishing. Many companies are conscious of the advantages, however there are a number of challenges delaying their implementation plans. That is the pure dwelling of open supply, offering the elements and basis, and creating an area for innovators to return collectively and share their concepts in a means that’s self–sustainable.

Extremely–low energy ML on the edge

Subsequent, Evgeni Gousev, senior director at Qualcomm, gave an of tinyML, which allows extremely–low energy machine studying (ML) on the very edge. In his discuss, he coated many points equivalent to tinyML fundamentals, its markets and values, and gave many examples. He mentioned latest developments from Qualcomm and offered details about the tinyML basis, ecosystem, initiatives and occasions, and academic actions.

Gousev additional outlined tinyML as ML architectures, methods, instruments, and approaches able to acting on–gadget analytics for a wide range of sensing modalities, equivalent to imaginative and prescient, audio, movement, chemical substances, and extra, concentrating on predominantly battery–operated gadgets.

The important thing tinyML progress drivers embrace extra environment friendly {hardware}, power environment friendly algorithms, extra mature software program infrastructure, instruments, numerous ecosystems, rising the variety of purposes, company funding, VC funding, and elevated begin–ups.

It’s predicted that one billion tinyML gadgets shall be shipped in 2024 and would method 5.4 billion in 2026. The expansion is in double digits. He described many helpful purposes of tinyML, equivalent to voice recognition, environmental sensors, predictive upkeep, gesture management, and augmented actuality.

AI/ML for Chip Design

Wrapping up the primary day, Venkat Thanvantri, VP of machine kearning R&D at Cadence introduced advances in AI/ML for chip design. The emergence of ML has unlocked many new purposes and reworked person experiences.

Electronics design automation (EDA) is an software space that delivers worth by offering automation and abstraction. ML know-how is having a equally transformative impression on EDA, accelerating execution of algorithms, bettering high quality of outcomes, and now considerably bettering the productiveness of customers. Specifically, he detailed using reinforcement studying to automate and optimize outcomes for the digital design and signoff movement.

Area particular architectures

On the second day of the symposium, Suk Hwan Lim government VP at Samsung, addressed advantageous grained area particular architectures for numerous workloads. He coated deep studying (DL) purposes and workloads, area particular neural processing models (NPU), and future instructions for NPUs. There are a number of DL purposes, together with speech recognition, voice activation, textual content to speech, authentication, picture classification, object detection, semantic segmentation, and picture processing.

Coaching is usually carried out within the cloud, whereas inference is deployed on the sting. The compute complexity varies by 5–7 orders of magnitude for these purposes. Thus, he categorized advantageous grained numerous processors for numerous purposes — micro NPU for audio/at all times–on purposes; normal NPUs for small spatial resolutions with deep community; and picture processing NPUs for big particular resolutions with shallow networks.

Vital enhancements in power, space, effectivity, and utilization are wanted and could be achieved by way of algorithms, compiler, structure, and circuits.

Reaching Greener AI with tinyML

Subsequent, Tune Han, assistant professor at MIT, described the function of tinyML and the way greener AI could be achieved. TinyML and environment friendly DL could make AI greener and simply deployable to the web of issues (IoT). AI purposes can generate excessive energy and have detrimental results on the atmosphere.

There’s a push for knowledge compression, pruning, and different methods to cut back computation and thereby energy. New fashions could be developed to enhance latency and accuracy. Handbook design is difficult, and automation is required. Han has proposed and developed a {hardware}–conscious neural community search referred to as “As soon as–for–All”.

Whereas computationally costly, As soon as–for–All requires coaching solely as soon as after which produces a number of fashions which can be utilized for inference. This method reduces the info utilization and ends in much less computation and, therefore, a decrease carbon footprint.

Additionally, sparse consideration and progressive quantization concepts are used to prune the tokens in pure language processing. Activation is the primary bottleneck and never the trainable parameter area. Activation minimization results in important reminiscence discount in IoT. All of the methods described in his discuss can assist AI to be greener.


Philip Wong, the Willard R. and Inez Kerr Bell professor within the College of Engineering at Stanford College, introduced “N3XT–3D–MOSAIC: Area–Particular Expertise for AI Compute”, noting that twenty first century purposes are going to be knowledge–centric. Knowledge analytics, ML, and AI purposes are going to dominate, from knowledge heart to cell and IoT, in addition to from accumulating and processing to curating knowledge and deriving info.

Many techniques might want to be taught and adapt on the fly. Three–dimensional integration is without doubt one of the main know-how instructions for built-in circuits.  Wong gave an outline of the brand new supplies and gadget applied sciences which will have to be developed to understand monolithic 3D integration with a number of logic transistor and reminiscence gadget layers. He gave two examples of compute–in–reminiscence chips that characteristic resistive switching random-access reminiscence integration with CMOS logic as an illustration of how future 3D techniques could also be designed.

Resistive switching random-access reminiscence (RRAM) array (Supply: Stanford College) (Click on picture to enlarge)

Reminiscence–compute subsystem

Steve Pawlowski, VP at Micron Expertise, talked about rethinking the reminiscence–compute subsystem. One of the best machines battle on workloads requiring increased reminiscence and community efficiency. The problem stays to enhance general techniques design beginning with reminiscence.

Pawlowski described evolving utilization fashions driving compute calls for that require huge quantities of information. He prompt present system bottlenecks will worsen until one thing is completed. New reminiscence–compute architectures can have the best impression on techniques. Nonetheless, new architectures will take time to evolve. However market pushed use circumstances can velocity adoption. The important thing to most power and efficiency is the combination of compute into reminiscence. New gadgets should construct from and evolve the dominant software program ecosystem. Many future neural networks would require increased bandwidths, which will increase power consumption. Reminiscence power is interconnect dominated. Reminiscence bandwidth can also be pin and locality dominated. A more in-depth coupling of reminiscence and compute is the trail ahead.

Ubiquitous AI

Subsequently, Pradeep Dubey, senior Fellow at Intel, coated the “Period of Ubiquitous AI”. AI is touching, if not remodeling, each side of our lives. AI is impacting not simply what computing can do for us, however quite how computing will get executed.

Quick–evolving AI algorithms are driving demand for normal–function computing that can’t be met by “enterprise as standard” engineering. On the similar time, programmers are sometimes knowledge scientists, not pc scientists; anticipating programmers to determine more and more advanced {hardware} on their very own simply doesn’t work.

Architects are due to this fact wanted greater than ever — chip architects to create new processors; techniques architects to design new knowledge facilities; software program architects to design new frameworks; and AI architects to churn out new fashions and new algorithms. Are we as much as the duty? Or do we have to increase human architects with AI to fulfill the problem?

Sustainable and accountable computing

team of researchers at the University of Cambridge in the U.K., led by Jenny Zhang from the Yusuf Hamied Department of Chemistry, successfully demonstrated the use of bacteria and photosynthesis for solar-based energy harvesting.

The symposium’s concluding discuss was given by Tamar Eilam, an IBM fellow. She introduced IBM’s initiative associated to sustainable and accountable computing. IBM historically has centered on privateness of information, safety, and ethics.

One other key consideration is lowering carbon footprint. For instance, some AI jobs eat a carbon footprint equal to that of the lifetime of 5 vehicles. It’s predicted that electrical energy utilization within the datacenter may improve by 8% by the yr 2030.

Carbon footprints are driving AI to maneuver to cloud. It’s essential to quantify carbon footprint power (CFE), which is outlined as product of IT tools power, energy utilization effectiveness (overhead energy conversion and cooling), and supply of the power (coal, nuclear, and so forth).

Organizations must report CFE to find scorching spots and optimize. To make greener AI, using renewable power, controlling the info utilization, and discovering alternate methods to cut back energy are important. With the intention to do that, workloads have to be monitored, enabling a breakdown of carbon footprints by cloud tenants and purposes.


The AI Compute Symposium additionally featured a poster session, organized into 3 parallel tracks. Out of 54 posters, there have been prime 3 finest posters awarded from every monitor. The listing of winners is given on the symposium web site.

The symposium closed with a panel dialogue on accountable computing with 4 panelists together with Tamar Eilam (IBM), Irena Risch (McGill), Evgeni Gousev (Qualcomm), and Bouchra Bouqata (Amazon). Panelists addressed questions spanning a spread of controversial subjects, together with the environmental impression of datacenter workloads, constructing belief in AI, eliminating bias in AI, and guaranteeing knowledge privateness.

Replays the of complete two–day symposium are additionally accessible on the symposium web site.

This report was compiled by Rajiv Joshi, normal chair; Arvind Kumar, program co–chair; and Matt Ziegler, program co–chair.



Please enter your comment!
Please enter your name here

Most Popular

Recent Comments