Showing posts with label AI. Show all posts
Showing posts with label AI. Show all posts

Monday, August 29, 2022

Cerebras opens engineering center in Bangalore

Cerebras Systems, which offers a Wafer Scale Engine AI processor, opened an engineering office in Bangalore, India. 

The new facility is led by industry veteran Lakshmi Ramachandran. Prior to joining Cerebras, she was with Intel in various engineering and leadership roles. Most recently, she was Senior Director at Intel's Data Center and AI group, responsible for delivering key capabilities of deep learning software for AI accelerators. She has extensive experience in scaling business operations and establishing technical engineering teams in India.

“India in general and Bangalore in particular, is extremely well-positioned to be a hotbed for AI innovation. It has world leading universities, pioneering research institutions and a large domestic enterprise market,” said Andrew Feldman, CEO and Co-Founder of Cerebras Systems. “Cerebras is committed to being a leader in this market. Under Lakshmi’s leadership, we are rapidly hiring top-notch engineering talent for Cerebras Systems India, as well as supporting sophisticated regional customers who are looking to do the most challenging AI work more quickly and easily.”



Thursday, July 28, 2022

Aruba advances its AIOps capabilities

Aruba introduced new AIOps capabilities for its Aruba ESP (Edge Services Platform) aimed at reducing the time spent on manual tasks such as network troubleshooting, performance tuning, and Zero Trust/SASE security enforcement. 

Aruba said AIOps, which it has been developing since 2013, is now being used not just network troubleshooting but also performance optimization and critical security controls.

The new capabilities are powered by Aruba’s massive data lake, which continuously and anonymously collects and analyzes device, user, and location data from over 120,000 Aruba Central customers, from more than 2 million network devices and 200 million clients per day. Aruba’s AI is trained by the high volume and wide variety of network and client data. 

“For AI results that customers can trust, the key ingredient is not a mathematical model, but access to a large volume and variety of data to train the models to produce reliable results across all network topologies. Without that foundation, so-called “AI” is nothing more than demoware,” said Larry Lunetta, vice president of portfolio solutions marketing at Aruba. “Fueled by our data lake, our AIOps solutions help enterprises reduce trouble tickets by up to 75 percent while optimizing their network performance by 25 percent or more.”

The new AI-powered IT efficiency features include:

  • Aruba Client Insights: Automatically identifies each endpoint connecting to the network with up to 99% accuracy, which is especially important as increasing numbers of IoT devices are added to networks, sometimes without approval from IT. This allows organizations to better understand what’s on their networks, automate access privileges, and monitor the behavior of each client’s traffic flows to more rapidly spot attacks and take action.
  • AI-powered Firmware Recommender: Provides IT teams with the best version of firmware to run for the wireless access points in their environments – regardless of model numbers. This greatly reduces support calls and guesswork that network admins face, and helps ensure new features and fixes are implemented more quickly.
  • AI Search in Spanish: The same built-in natural language search function in Aruba Central shows its versatility by now supporting queries and responses in Spanish to satisfy the needs of our second largest geographical user community.
  • Automated Infrastructure Predictions: Leverages Aruba’s AI Assist feature and Aruba Support outreach to recognize possible hardware and software infrastructure issues for preemptive engagement that can consist of firmware upgrades or recommended hardware replacement.

https://news.arubanetworks.com/news-release-details/2022/Aruba-Helps-Network-Teams-Overcome-Scarce-Staff-Resources-with-First-AIOps-Solution-that-Combines-Network-and-Security-Insights-for-Improved-IT-Efficiency/default.aspx



Video: Tipping point for A.I. in managing networks?



Eventually all networks will be fully managed by artificial intelligence, but that's not the current reality -- humans are running the show.

How close are we to the tipping point where A.I. can be trusted to make critical decisions for network operations? And what will it take to earn our trust?

Here is a perspective from Larry Lunetta, VP WLAN & Security Marketing at Aruba, a Hewlett Packard Enterprise company.

https://youtu.be/ozNa0FpJQHw

Tuesday, July 12, 2022

Intel releases AI reference kits

Intel released the first set of open source AI reference kits, including AI model code, end-to-end machine learning pipeline instructions, libraries and Intel oneAPI components, for on-prem, cloud and edge environments. 

Intel said its kits enable data scientists and developers to learn how to deploy AI faster and more easily across healthcare, manufacturing, retail and other industries with higher accuracy, better performance and lower total cost of implementation.

Four kits are available:

  • Utility asset health: This predictive analytics model was trained to help utilities deliver higher service reliability. It uses Intel-optimized XGBoost through the Intel oneAPI Data Analytics Library to model the health of utility poles with 34 attributes and more than 10 million data points. 
  • Visual quality control: The AI Visual QC model was trained using Intel AI Analytics Toolkit, including Intel Optimization for PyTorch and Intel Distribution of OpenVINO toolkit, both powered by oneAPI to optimize training and inferencing to be 20% and 55% faster, respectively, compared to stock implementation of Accenture visual quality control kit without Intel optimizations2 for computer vision workloads across CPU, GPU and other accelerator-based architectures. Using computer vision and SqueezeNet classification, the AI Visual QC model used hyperparameter tuning and optimization to detect pharmaceutical pill defects with 95% accuracy.
  • Customer chatbot: Conversational chatbots have become a critical service to support initiatives across the enterprise. AI models that support conversational chatbot interactions are massive and highly complex. This reference kit includes deep learning natural language processing models for intent classification and named-entity recognition using BERT and PyTorch. Intel Extension for PyTorch and Intel Distribution of OpenVINO toolkit optimize the model for better performance – 45% faster inferencing compared to stock implementation of Accenture customer chatbot kit without Intel optimizations3 – across heterogeneous architectures, and allow developers to reuse model development code with minimal code changes for training and inferencing.
  • Intelligent document indexing: Enterprises process and analyze millions of documents every year, and many of the semi-structured and unstructured documents are routed manually. AI can automate the processing and categorizing of these documents for faster routing and lower manual labor costs. Using a support vector classification (SVC) model, this kit was optimized with Intel Distribution of Modin and Intel Extension for Scikit-learn powered by oneAPI.

https://www.intel.com/content/www/us/en/developer/topic-technology/artificial-intelligence/reference-kit.html

Intel to acquire real-time continuous optimization software

Intel agreed to acquire Granulate Cloud Solutions Ltd., an Israel-based developer of real-time continuous optimization software. Financial terms were not disclosed.Granulate’s autonomous optimization service helps reduce CPU utilization and application latencies. It does this by learning the customer’s application and deploying a customized set of continuous optimizations at runtime. This enables deployment on smaller compute clusters and instance...

Intel and Microsoft contribute Scalable I/O Virtualization spec to OCP

 Intel and Microsoft have contributed a Scalable I/O Virtualization (SIOV) specification to the Open Compute Project (OCP), enabling virtualization of PCI Express and Compute Express Link devices in cloud servers. SIOV is hardware-assisted I/O virtualization with the potential to support thousands of virtualized workloads per server. SIOV moves the non-performance-critical virtualization and management logic off the PCIe device and into...


Wednesday, June 22, 2022

Cerebras sets a new record for largest AI model

Cerebras Systems announced an ability to train models with up to 20 billion parameters on a single CS-2 system.

The Cerebras WSE-2 is the largest processor ever built, boasting 2.55 trillion more transistors and 100 times as many compute cores as the largest GPU. 

By enabling a single CS-2 to train these models, Cerebras reduces the system engineering time necessary to run large natural language processing (NLP) models from months to minutes. It also eliminates one of the most painful aspects of NLP — namely the partitioning of the model across hundreds or thousands of small graphics processing units (GPU).

“In NLP, bigger models are shown to be more accurate. But traditionally, only a very select few companies had the resources and expertise necessary to do the painstaking work of breaking up these large models and spreading them across hundreds or thousands of graphics processing units,” said Andrew Feldman, CEO and Co-Founder of Cerebras Systems. “As a result, only very few companies could train large NLP models – it was too expensive, time-consuming and inaccessible for the rest of the industry. Today we are proud to democratize access to GPT-3XL 1.3B, GPT-J 6B, GPT-3 13B and GPT-NeoX 20B, enabling the entire AI ecosystem to set up large models in minutes and train them on a single CS-2.”



Thursday, May 26, 2022

Meta selects Azure as strategic cloud provider for AI research

Meta has selected Azure as a strategic cloud provider to help accelerate AI research and development. 

Specifically, Meta will utilize a dedicated Azure cluster of 5400 GPUs using the latest virtual machine (VM) series in Azure (NDm A100 v4 series, featuring NVIDIA A100 Tensor Core 80GB GPUs) for some of their large-scale AI research workloads.

In addition, the companies agreed to collaborate to scale PyTorch adoption on Azure. 

“We are excited to deepen our collaboration with Azure to advance Meta’s AI research, innovation, and open-source efforts in a way that benefits more developers around the world,” Jerome Pesenti, Vice President of AI, Meta. “With Azure’s compute power and 1.6 TB/s of interconnect bandwidth per VM we are able to accelerate our ever-growing training demands to better accommodate larger and more innovative AI models. Additionally, we’re happy to work with Microsoft in extending our experience to their customers using PyTorch in their journey from research to production.”

https://azure.microsoft.com/en-us/blog/meta-selects-azure-as-strategic-cloud-provider-to-advance-ai-innovation-and-deepen-pytorch-collaboration/

Thursday, March 10, 2022

SambaNova expands its exec team

SambaNova Systems, a start-up developing a software, hardware, and solutions platform to run AI and Deep Learning applications, announces the appointment of Peter Buckingham as Vice President of Software Engineering. 

Buckingham has held leadership positions at VMware, Bitfusion, Sun Microsystems, Dell and Waypoint.

“With AI, we are experiencing one of the biggest transitions in computing history since the internet and the transition is pushing the boundaries of what’s possible now. Enterprises are rapidly deploying AI to accelerate their transformation and jump into the future of computing,” said Rodrigo Liang, CEO of SambaNova Systems. “I’m excited to welcome Peter to the team. His experience will enable SambaNova to quickly create software to help enterprise leaders unleash AI’s power.”

“I’m honored to join SambaNova’s leadership team. This is an incredible opportunity to create first-to-market software that accelerates AI adoption across the globe,” said Peter Buckingham, VP of Software at SambaNova. “Seventy-two percent of enterprises struggle with AI adoption yet they desperately need to deploy it today. The software we’re creating at SambaNova will help them do that.”

SambaNova is based in Palo Alto, California.

SambaNova raises $676 million for its AI platform

SambaNova Systems, a start-up based in Palo Alto, California, announced $676 million in Series D funding for its software, hardware and services to run AI applications.SambaNova’s flagship offering is Dataflow-as-a-Service (DaaS), a subscription-based, extensible AI services platform designed to jump-start enterprise-level AI initiatives, augmenting organizations’ AI capabilities and accelerating the work of existing data centers, allowing the organization...


Monday, February 7, 2022

Esperanto looks to IFS to fab its RISC-V-based AI accelerator chips

Esperanto Technologies, a start-up based in Mountain View, California, confirmed that it will use Intel Foundry Services silicon and chiplet packaging technologies to advance its RISC-V-based technology and deliver its massively parallel AI acceleration silicon solutions spanning from cloud to edge.

Esperanto Technologies is developing massively-parallel 64-bit RISC-V-based Tensor compute cores currently delivered in the form of a single chip with 1088 ET-Minion compute cores and a high-performance memory system. Designed to meet the high-performance and low-power requirements of large-scale datacenter customers, Esperanto’s existing RISC-V-based inference chip is a general purpose, parallel processing solution that can accelerate many parallelizable workloads. It is designed to run any machine learning (ML) workload well, and to excel at ML recommendation models, one of the most important types of AI workloads in many large datacenters.

“Intel Foundry Services is excited to add Esperanto’s massively-parallel AI accelerators to the IFS ecosystem,” said Bob Brennan, Vice unity are invaluable in helping to advance and proliferate RISC-V solutions from Esperanto and others, ultimately benefiting all end users,” said Art Swift, president and CEO at Esperanto Technologies.

https://www.esperanto.ai/technology/



Thursday, December 9, 2021

Expedera raises $18M for deep learning accelerator from founders of Marvell

Expedera, a start-up based in Santa Clara, California announced a $18 million Series A funding round led by Dr. Sehat Sutardja and Weili Dai (founders of Marvell Technology Group) and other prominent semiconductor industry investors. 

Expedera is developing a deep learning accelerator IP that is scalable up to 128 TOPS with a single core and to PetaOps with multi-core. The company says it will achieve the highest performance per watt. The solution is aimed at a wide range of AI inference applications, particularly at the edge. Expedera’s Origin IP and software platform supports popular AI frontends including TensorFlow, ONNX, Keras, Mxnet, Darknet, CoreML and Caffe2 through Apache TVM. By licensing its technology as a semiconductor IP, Expedera enables any chip designer to add state-of-the-art AI functionality to their product.

The latest funding brings the total amount raised to over $27 million.

“This financing underscores the success that Expedera has had so far and will enable us to expand our portfolio and team to meet the market needs,” said Da Chuang, CEO of Expedera. “We are incredibly happy to have Weili Dai and Sehat Sutardja lead this round. As highly respected veterans of the semiconductor industry, they have a unique understanding of the market and customer needs. I look forward to a long partnership.”

“Device makers have typically needed to build their own chips and usually, only the largest companies could afford to do so,” said Mr. Gwennap. “Expedera’s IP model provides a more cost effective way to address the sprawling edge AI market. A single IP supplier can license to any or all of the numerous chip vendors that supply a multitude of device makers in the edge market.”

http://www.expedera.com



Tuesday, November 2, 2021

Juniper adds Wi-Fi 6E APs powered by its Mist AI

Juniper Networks introduced two new 6 GHz access points that leverage its Mist AI to maximize Wi-Fi performance and capacity while simplifying IT operations. In addition, Juniper is introducing a new IoT Assurance service that streamlines and scales the onboarding and securing of IoT devices without Network Access Control (NAC). These enhancements to the Juniper wireless access portfolio further the company’s experience-first networking mission so that IT administrators can deliver the best network experiences to their end users.

“While it is relatively easy to support new standards like Wi-Fi 6E in hardware and to onboard a small number of IoT devices using pre-shared keys, Juniper has again put customers first by applying the proven benefits of Mist AI and the modern microservices cloud to these environments to deliver unparalleled performance, agility, ease and scale,” said Jeff Aaron, VP Enterprise Marketing. “From Day Zero installation through ongoing monitoring, management and troubleshooting of the network, Juniper continues to stand out for an experience-first approach to networking that delivers the best experiences for operators and end users from the client to the cloud.”

The two new tri-band access points (APs), managed via the same Mist cloud and AI engine as the rest of the Juniper Mist portfolio, include:

  • AP 45 – 2.4 GHz/5 GHz/6 GHz quad-radio, 4x4:4SS, vBLE array
  • AP 34 – 2.4 GHz/5 GHz/6 GHz quad-radio, 2x2:2SS, omni BLE

The new access points complement the existing 2.4- and 5-GHz APs currently in the Juniper Mist portfolio, which support both 802.11ac and 802.11ax protocols. All APs incorporate either Juniper’s patented virtualized Bluetooth LE or omnidirectional BLE antenna in addition to Wi-Fi to provide a wide range of location-based services in a scalable and cost effective manner.



Wednesday, October 13, 2021

Hailo raises $136 million for its edge AI processors

Hailo, a start-up based in Tel Aviv, raised $136 million in a Series C round of funding for its edge processor designed for AI workloads.

The Hailo-8 edge AI processor boasts up to 26 tera-operations per second (TOPS) performance, capable of processing of FHD stream in real-time, and with typical power consumption of 2.5W, according to the company.

The funding round was led by Poalim Equity and Gil Agmon with participation from existing investors including Hailo Chairman Zohar Zisapel, ABB Technology Ventures (ATV), Latitude Ventures and OurCrowd; and new investors Carasso Motors, Comasco, Shlomo Group, Talcar Corporation Ltd. and Automotive Equipment (AEV).

Hailo was established in Israel in 2017 by members of the Israel Defense Forces’ elite technology unit.

https://hailo.ai/

Tuesday, September 14, 2021

Juniper enhances its AI-driven enterprise portfolio

Juniper Networks announced enhancements to the Juniper Mist cloud and AI engine, which include EVPN-VXLAN campus fabric management and additional Marvis Actions for proactive problem remediation. The new features within the AI-driven enterprise portfolio that enable customers to scale and simplify the rollout of their campus wired and wireless networks while bringing greater insight and automation to network operators.


The latest additions to Juniper’s AI-driven enterprise include:

AI-driven campus fabric management via the Juniper Mist Cloud: By enabling EVPN-VXLAN campus fabric management via the Juniper Mist Wired Assurance cloud service, Juniper offers the ability to simplify wired, wireless and WAN via a common cloud and AIOps engine. Juniper EVPN-VXLAN fabric leverages the same Juniper platform used to manage wired access in the campus, Juniper Mist Wired Assurance and the Marvis Virtual Network Assistant (VNA), bringing IT administrators unparalleled automation, insight and troubleshooting. The Juniper Mist Cloud empowers administrators to choose a topology, define networks of interest, identify required physical connections and apply the correct underlying policies in a seamless fashion. In addition, the Juniper Mist solution enables customers to leverage a common operational schema across LAN, WLAN and WAN environments, a key part of the Juniper client-to-cloud differentiation.

New Marvis Actions that provide deeper insight for faster problem remediation: Marvis Actions takes insight derived from the Mist AI engine, such as the root cause of a problem, and delivers actionable recommendations for IT managers via a simple dashboard. Additional actions have been added to the Marvis VNA software subscription to detect and correct even more wired/wireless/WAN issues, such as persistently failing wired/wireless clients, bad cables, access point (AP) coverage holes, bad WAN links and insufficient RF capacity, among others.

“Juniper is committed to Experience-First Networking, where our enterprise solutions leverage proactive automation, assured user experiences, agile cloud services and connected security to deliver the best end-user and operator experiences from client-to-cloud,” said Jeff Aaron, Vice President of Enterprise Marketing at Juniper. “In the campus and branch, this means leveraging AIOps, driven by Mist AI and the cloud to maximize user productivity and efficiency while minimizing IT costs through simplified operations and prescriptive insights. Today, we are excited to accelerate in these areas – and further distance ourselves from the competition – with new enhancements to our Juniper Mist Wired Assurance, Wireless Assurance and Marvis Virtual Network Assistant (VNA) cloud services that make it even easier to deploy, operate and troubleshoot campus networks at scale.”

Tuesday, August 24, 2021

Cerebras advances its "Brain-scale AI"

Cerebras Systems disclosed progress in its mission to deliver a "brain-scale" AI solution capable of supporting neural network models of over 120 trillion parameters in size. 

Cerebras’ new technology portfolio contains four innovations: Cerebras Weight Streaming, a new software execution architecture; Cerebras MemoryX, a memory extension technology; Cerebras SwarmX, a high-performance interconnect fabric technology; and Selectable Sparsity, a dynamic sparsity harvesting technology.

  • Cerebras Weight Streaming enables the ability to store model parameters off-chip while delivering the same training and inference performance as if they were on chip. This new execution model disaggregates compute and parameter storage – allowing researchers to flexibly scale size and speed independently – and eliminates the latency and memory bandwidth issues that challenge large clusters of small processors. It is designed to scale from 1 to up to 192 CS-2s with no software changes.
  • Cerebras MemoryX is a memory extension technology. MemoryX will provide the second-generation Cerebras Wafer Scale Engine (WSE-2) up to 2.4 Petabytes of high performance memory, all of which behaves as if it were on-chip. With MemoryX, CS-2 can support models with up to 120 trillion parameters.
  • Cerebras SwarmX is a high-performance, AI-optimized communication fabric that extends the Cerebras Swarm on-chip fabric to off-chip. SwarmX is designed to enable Cerebras to connect up to 163 million AI optimized cores across up to 192 CS-2s, working in concert to train a single neural network.
  • Selectable Sparsity enables users to select the level of weight sparsity in their model and provides a direct reduction in FLOPs and time-to-solution. Weight sparsity is an exciting area of ML research that has been challenging to study as it is extremely inefficient on graphics processing units. Selectable sparsity enables the CS-2 to accelerate work and use every available type of sparsity—including unstructured and dynamic weight sparsity—to produce answers in less time.

“Today, Cerebras moved the industry forward by increasing the size of the largest networks possible by 100 times,” said Andrew Feldman, CEO and co-founder of Cerebras. “Larger networks, such as GPT-3, have already transformed the natural language processing (NLP) landscape, making possible what was previously unimaginable. The industry is moving past 1 trillion parameter models, and we are extending that boundary by two orders of magnitude, enabling brain-scale neural networks with 120 trillion parameters.”

https://cerebras.net/news/cerebras-systems-announces-worlds-first-brain-scale-artificial-intelligence-solution/


Cerebras unveils 2nd-gen, 7nm Wafer Scale Engine chip

Cerebras Systems introduced its Wafer Scale Engine 2 (WSE-2) AI processor, boasting 2.6 trillion transistors and 850,000 AI optimized cores.

The wafer-sized processor, which is manufactured by TSMC on its 7nm-node, more than doubles all performance characteristics on the chip - the transistor count, core count, memory, memory bandwidth and fabric bandwidth - over the first generation WSE. 

“Less than two years ago, Cerebras revolutionized the industry with the introduction of WSE, the world’s first wafer scale processor,” said Dhiraj Mallik, Vice President Hardware Engineering, Cerebras Systems. “In AI compute, big chips are king, as they process information more quickly, producing answers in less time – and time is the enemy of progress in AI. The WSE-2 solves this major challenge as the industry’s fastest and largest AI processor ever made.”


The processors powers the Cerebras CS-2 system, which the company says delivers hundreds or thousands of times more performance than legacy alternatives, replacing clusters of hundreds or thousands of graphics processing units (GPUs) that consume dozens of racks, use hundreds of kilowatts of power, and take months to configure and program. The CS-2 fits in one-third of a standard data center rack.

Early deployment sites for the first generation Cerebras WSE and CS-1 included Argonne National Laboratory, Lawrence Livermore National Laboratory, Pittsburgh Supercomputing Center (PSC) for its groundbreaking Neocortex AI supercomputer, EPCC, the supercomputing centre at the University of Edinburgh, pharmaceutical leader GlaxoSmithKline, and Tokyo Electron Devices, amongst others.

“At GSK we are applying machine learning to make better predictions in drug discovery, so we are amassing data – faster than ever before – to help better understand disease and increase success rates,” said Kim Branson, SVP, AI/ML, GlaxoSmithKline. “Last year we generated more data in three months than in our entire 300-year history. With the Cerebras CS-1, we have been able to increase the complexity of the encoder models that we can generate, while decreasing their training time by 80x. We eagerly await the delivery of the CS-2 with its improved capabilities so we can further accelerate our AI efforts and, ultimately, help more patients.”

“As an early customer of Cerebras solutions, we have experienced performance gains that have greatly accelerated our scientific and medical AI research,” said Rick Stevens, Argonne National Laboratory Associate Laboratory Director for Computing, Environment and Life Sciences. “The CS-1 allowed us to reduce the experiment turnaround time on our cancer prediction models by 300x over initial estimates, ultimately enabling us to explore questions that previously would have taken years, in mere months. We look forward to seeing what the CS-2 will be able to do with more than double that performance.”

https://cerebras.net/product

Tuesday, July 20, 2021

Untether AI raises $125 million for high-performance silicon

Untether AI, a start-up based in Toronto, announced an oversubscribed $125 million funding round for its at-memory computation and AI inference acceleration silicon.

The latest funding was led by an affiliate of Tracker Capital Management and by Intel Capital, with participation from new investor Canada Pension Plan Investment Board (“CPP Investments”) and existing investor Radical Ventures.

As part of the funding round, Tracker Capital Senior Advisor Dr. Shaygan Kheradpir will join Untether AI’s Board of Directors. Previously, Dr. Kheradpir served as Verizon’s Group CIO, Barclays Bank Group COO, and CEO of Coriant and Juniper Networks.

“I am pleased to add Tracker Capital to our prestigious group of investors and welcome Shaygan to our Board,” said Arun Iyengar, CEO, Untether AI. “Tracker Capital’s unmatched experience and relationships across sectors will help speed our engagements in multiple high-value markets, including telecom, technology, financial services, retail, and defense. I am also thrilled to welcome CPP Investments to the Untether AI family. With the new funding round and partnerships, we will be able to expand our current product reach and accelerate the development of our next generation products."

Saf Yeboah-Amankwah, Intel Chief Strategy Officer added: “We have been an investor in Untether AI since the seed round. During that time Untether AI has assembled a world-class management team, developed and launched an exceptional product, and is now poised for growth in the burgeoning AI inference acceleration space.”



The dramatic increase in the usage of AI, along with its heavy computational requirements, is overwhelming traditional compute architectures, and drastically increasing power consumption in datacenters. 


Untether AI said its at-memory compute architecture and tsunAImi accelerator cards can achieve record-breaking energy efficiency and compute density for inference acceleration. 


www.untether.ai

Monday, June 21, 2021

HPE acquires Determined AI

Hewlett Packard Enterprise has acquired Determined AI, a start-up based in San Francisco with software stack to train AI models faster using its open source machine learning (ML) platform. Financial terms were not disclosed.

Determined AI was founded in 2017 by Neil Conway, Evan Sparks, and Ameet Talwalkar. It launched its open-source platform in 2020. 

HPE will combine Determined AI’s unique software solution with its AI and high performance computing (HPC) offerings to enable ML engineers to easily implement and train machine learning models to provide faster and more accurate insights from their data in almost every industry.  

“As we enter the Age of Insight, our customers recognize the need to add machine learning to deliver better and faster answers from their data,” said Justin Hotard, senior vice president and general manager, HPC and Mission Critical Solutions (MCS), HPE. “AI-powered technologies will play an increasingly critical role in turning data into readily available, actionable information to fuel this new era. Determined AI’s unique open source platform allows ML engineers to build models faster and deliver business value sooner without having to worry about the underlying infrastructure. I am pleased to welcome the world-class Determined AI team, who share our vision to make AI more accessible for our customers and users, into the HPE family.”

“The Determined AI team is excited to join HPE, who shares our vision to realize the potential of AI,” said Evan Sparks, CEO of Determined AI. “Over the last several years, building AI applications has become extremely compute, data, and communication intensive. By combining with HPE’s industry-leading HPC and AI solutions, we can accelerate our mission to build cutting edge AI applications and significantly expand our customer reach.” 


Wednesday, June 9, 2021

Xilinx debuts Versal AI Edge series processors

Xilinx introduced the Versal AI Edge series processors, boasting 4X the AI performance-per-watt versus GPUs and 10X greater compute density versus previous-generation adaptive SoCs.

Xilinx is positioning the new Versal AI Edge adaptive compute acceleration platforms (ACAPs) for a range of applications including: automated driving with the highest levels of functional safety, collaborative robotics, predictive factory and healthcare systems, and multi-mission payloads for the aerospace and defense markets. The portfolio features AI Engine-ML to deliver 4X machine learning compute compared to the previous AI Engine architecture and integrates new accelerator RAM with an enhanced memory hierarchy for evolving AI algorithms. These architectural innovations deliver up to 4X AI performance-per-watt versus GPUs and lower latency resulting in far more capable devices at the edge.

"Edge computing applications require an architecture that can evolve to address new requirements and scenarios with a blend of flexible compute processing within tight thermal and latency constraints,” said Sumit Shah, senior director, Product Management and Marketing at Xilinx. “The Versal AI Edge series delivers these key attributes for a wide range of applications requiring greater intelligence, making it a critical addition to the Versal portfolio with devices that scale from intelligent edge sensors to CPU accelerators.”

The Versal AI Edge series takes the production-proven 7nm Versal architecture and miniaturizes it for AI compute at low latency, all with power efficiency as low as six watts and safety and security measures required in edge applications. As a heterogeneous platform with diverse processors, the Versal AI Edge series matches the engine to the algorithm, with Scalar Engines for embedded compute, Adaptable Engines for sensor fusion and hardware adaptability, and Intelligent Engines for AI inference that scales up to 479 (INT4) TOPS2—unmatched by ASSPs and GPUs targeting edge applications—and for advanced signal processing workloads for vision, radar, LiDAR, and software defined radio.

Sampling is available to early access customers, with shipments expected during the first half of 2022.

https://www.xilinx.com/versal-ai-edge

Thursday, April 29, 2021

Vectra AI raises $130 million for automated threat detection/response

Vectra AI, a start-up based in San Jose, California, announced $130 million in new funding for its work in automated cyber threat detection and response. The company's mission is "to see and stop threats before they become breaches."

“Over the past year, we have witnessed a continuous series of the most impactful and widespread cyberattacks in history. To protect their employees and digital assets, our customers require security solutions that are smarter than today’s adversaries and provide coverage for cloud, data centers and SaaS applications” said Hitesh Sheth, president and chief executive officer at Vectra. “As we look to the future, Blackstone’s global presence, operational resources, and in-house technology expertise will help us achieve our mission to become one of the dominant cybersecurity companies in the world.”

The new $130 funding round was led by funds managed by Blackstone Growth. This brings Vectra's total funding since inception to more than $350 million at a post-money $1.2 billion valuation.

Viral Patel, a Senior Managing Director at Blackstone, said: “Vectra has a proven ability to stop in-progress attacks in the cloud, on corporate networks, and in private data centers for some of the top organizations in the world. The company has experienced extraordinary success through its commitment to combining innovative AI technology, first-class customer service, and top talent, and Blackstone is excited to become part of the Vectra team.”

For 2020, the Vectra reported a compound annual growth rate (CAGR) exceeding 100 percent, while sales of its Cognito Detect product for Microsoft Office 365 have grown at a rate of over 700 percent. 

http://www.vectra.ai

  • Vectra AI is headed by Hitesh Sheth (president and CEO), who previously was chief operating officer at Aruba Networks. Hitesh joined Aruba from Juniper Networks, where he was EVP/GM for its switching business and before that, SVP for the Service Layer Technologies group, which included security. Prior to Juniper, Hitesh held a number of senior management positions at Cisco.

Thursday, April 22, 2021

Expedera develops deep-learning accelerator (DLA) for AI silicon

Expedera, a start-up based in Santa Clara, California, emerged from stealth to unveil its Origin neural engine intellectual property (IP) for edge system silicon.

Expedera, which plans to license its deep-learning accelerator (DLA) technology to SoC designers, is targetting low-power edge devices like smartphones, tablets, computers, edge servers, and automotive. The company says its deep-learning accelerator (DLA) provides up to 18 TOPS/W at 7nm, which is up to ten times more than competitive offerings while minimizing memory requirements. Origin accelerates the performance of neural network models such as object detection, recognition, segmentation, super-resolution, and natural language processing. 

Expedera's co-founders include Da Chuang (CEO), who previously was cofounder and COO of Memoir Systems (an optimized memory IP startup acquired by Cisco); Siyad Ma (VP Engineering), who previously led Algorithmic TCAM ASIC and IP teams for Cisco Nexus7k, MDS, Cat4k/6k; and Sharad Chole (Chief Scientist), who previously was an architect at Cisco, Memoir Systems (Cisco), and Microsoft. 

https://www.expedera.com/

Tuesday, April 13, 2021

SambaNova raises $676 million for its AI platform

SambaNova Systems, a start-up based in Palo Alto, California, announced $676 million in Series D funding for its software, hardware and services to run AI applications.

SambaNova’s flagship offering is Dataflow-as-a-Service (DaaS), a subscription-based, extensible AI services platform designed to jump-start enterprise-level AI initiatives, augmenting organizations’ AI capabilities and accelerating the work of existing data centers, allowing the organization to focus on its business objectives instead of infrastructure.

At the core of DaaS is SambaNova’s DataScale, an integrated software and hardware systems platform with optimized algorithms and next-generation processors delivering unmatched capabilities and efficiency across applications for training, inference, data analytics, and high-performance computing. SambaNova’s software-defined-hardware approach has set world records in AI performance, accuracy, scale, and ease of use.

The funding round was led by SoftBank Vision Fund 2, and included additional new investors Temasek and GIC, plus existing backers including funds and accounts managed by BlackRock, Intel Capital, GV (formerly Google Ventures), Walden International and WRVI. This Series D brings SambaNova’s total funding to more than $1 billion and rockets its valuation to more than $5 billion.

SambaNova says it is working "to shatter the computational limits of AI hardware and software currently on the market — all while making AI solutions for private and public sectors more accessible."

“We’re here to revolutionize the AI market, and this round greatly accelerates that mission,” said Rodrigo Liang, SambaNova co-founder and CEO. “Traditional CPU and GPU architectures have reached their computational limits. To truly unleash AI’s potential to solve humanity’s greatest technology challenges, a new approach is needed. We’ve figured out that approach, and it’s exciting to see a wealth of prudent investors validate that.”

Stanford Professors Kunle Olukotun and Chris Ré, along with Liang, founded SambaNova in 2017 and came out of stealth in December 2020. Olukotun is known as the “father of the multi-core processor” and the leader of the Stanford Hydra Chip Multiprocessor (CMP) research project. Ré is an associate professor in the Department of Computer Science at Stanford University. He is a MacArthur Genius Award recipient, and is affiliated with the Statistical Machine Learning Group, Pervasive Parallelism Lab, and Stanford AI Lab.

http://www.sambanova.ai


Cerebras appoints CMO as it continues to grow

Cerebras Systems, a start-up developing a Wafer Scale Engine (WSE) chip that contains 1.2 trillion transistors, covers more than 46,225 square millimeters of silicon and contains 400,000 AI optimized compute cores, announced the appointment of Rupal Shah Hollenbeck as Vice President and Chief Marketing Officer (CMO). 

Prior to Cerebras, Hollenbeck served as senior vice president and CMO at Oracle, where she led the marketing transformation strategy for the company, while overseeing global brand and demand generation for all product areas. Previously, she held various senior leadership positions at Intel for more than two decades, most recently serving as Corporate Vice President and General Manager for Sales & Marketing in Intel’s Data Center division. Hollenbeck serves as an Independent Director of Check Point Software Technologies and is a Founding Limited Partner in Neythri Futures Fund, a venture fund dedicated to bringing diversity to the investment ecosystem.

“I am thrilled to join Cerebras’ industry-leading team as they tackle some of society’s most urgent and challenging problems with their groundbreaking CS-1 AI supercomputer,” said Hollenbeck. “I’ve been impressed with Cerebras’ customer traction over the past year, and I look forward to further accelerating this momentum with new global partnerships and customer deployments.”

Over the past year, Cerebras opened new offices in Tokyo, Japan and Toronto, Canada. The company also announced a series of wins for its flagship product, the CS-1, including deployments at Argonne National Laboratory, Lawrence Livermore National Laboratory, Pittsburgh Supercomputing Center (PSC) for its groundbreaking Neocortex AI supercomputer, EPCC, the supercomputing centre at the University of Edinburgh, and pharmaceutical leader GlaxoSmithKline.