Weka provides cloud storage solutions. It offers a shared parallel file system and storage infrastructure. It helps to tackle demanding storage performance challenges in data-intensive technical computing environments to solve problems. It accelerates time-to-insight from data and helps customers get high-powered information technology investments. It was formerly known as WekaIO. The company was founded in 2013 and is based in Campbell, California.
Research containing Weka
Get data-driven expert analysis from the CB Insights Intelligence Unit.
CB Insights Intelligence Analysts have mentioned Weka in 1 CB Insights research brief, most recently on Apr 14, 2023.
Expert Collections containing Weka
Expert Collections are analyst-curated lists that highlight the companies you need to know in the most important technology spaces.
Weka is included in 1 Expert Collection, including Artificial Intelligence.
Companies developing artificial intelligence solutions, including cross-industry applications, industry-specific products, and AI infrastructure solutions.
Weka has filed 92 patents.
Computer storage devices, Data management, IBM mainframe operating systems, Computer data storage, USB
Computer storage devices, Data management, IBM mainframe operating systems, Computer data storage, USB
Latest Weka News
Sep 19, 2023
MLCommons , the machine learning standards engineering consortium, has produced a machine learning storage benchmark with results from DDN, Nutanix, Weka and others. MLCommons was founded in 2020 by the people who produced the MLPerf benchmark for testing ML hardware performance in 2018. It wants to increase the AI/ML adoption rate by developing quality and performance measures, large-scale open datasets and common development practices and resources. MLCommons has more than 50 members including software startups, university researchers, cloud computing and semiconductor giants. Among them are Dell EMC, HPE, Huawei, Intel, Lenovo, Meta, Nvidia, Nutanix, and VMware. It has announced results from its MLPerf Inference v3.1 and first ever MLPerf Storage v0.5 benchmarks. David Kanter, executive director of MLCommons, stated: “Submitting to MLPerf is not trivial. It’s a significant accomplishment, as this is not a simple point and click benchmark. It requires real engineering work and is a testament to our submitters’ commitment to AI, to their customers, and to ML.” The MLCommons MLPerf Inference benchmark suite measures how fast systems can run models in a variety of deployment scenarios. The open source MLPerf Storage benchmark suite measures the performance of storage systems in the context of ML training workloads. It’s built on the codebase of DLIO – a benchmark designed for I/O measurement in high performance computing, adapted to meet current storage needs. Oana Balmau, Storage Working Group co-chair, stated: “Our first benchmark has over 28 performance results from five companies which is a great start given this is the first submission round. We’d like to congratulate MLPerf Storage submitters: Argonne National Laboratory (ANL using HPE ClusterStor), DDN, Micron, Nutanix, WEKA for their outstanding results and accomplishments.” There are both open and closed MLPerf Storage submissions. Closed submissions use the same reference model to ensure a level playing field across systems, while participants in the open division are permitted to submit a variety of models. We only look at the closed submissions as they enable cross-vendor comparisons in principle. The benchmark was created through a collaboration spanning more than a dozen industry and academic organizations and includes a variety of storage setups, including parallel file systems, local storage, and software defined storage. MLCommons says that when developing the next generation of ML models, it is a challenge to find the right balance between and efficient utilization of storage and compute resources. The MLPerf Storage benchmark helps overcome this problem by accurately modeling the I/O patterns posed by ML workloads, providing the flexibility to mix and match different storage systems with different accelerator types. The MLPerf Storage benchmark is intended be an effective tool for purchasing, configuring, and optimizing storage for machine learning applications, as well as for designing next-generation systems and technologies. Submitting systems are scored for two MLworkloads – medical image segmentation and natural language processing – with samples per second and MBps ratings in each category. MLCommons says samples/second is a metric that should be intuitively valuable to AI/ML practitioners, and the MBps metric should be intuitively valuable to storage practitioners. We look at MBps. The dataset used in each benchmark submission is automatically scaled to a size that prevents significant caching in the systems actually running the benchmark code. The submitted systems all use a simulated V100 accelerator (Nvidia V100 Tensor core GPU) with varying numbers of these accelerators tested, making cross-supplier comparisons difficult. There are GitHub reference links for details of code per submitter and system. We have extracted the basic systems and scores data from the full benchmark table. Then we have added a column of our own – MBps per accelerator – to try to normalize performance per accelerator and so enable cross-supplier comparisons: ALCF refers to the ALCF Discretionary Allocation Program, and shows results for HPE ClusterStor systems that are available in the Argonne National Laboratory Discretionary Allocation Program. Overall there is not that much difference between the suppliers and systems, with the image segmentation MBps/Accelerator scores bunched between 410 and 361. Micron has both the the highest score at 410, and the lowest, at 361. The natural language processing MBps numbers are much lower, with DDN having the single highest result, 1.7, and all the others varying between 0.12 and 0.10. We suggest it’s good to have a vendor-independent benchmark in this area at all, even if its first iteration is not that informative. We have no means of relating these numbers to supplier performance when delivering data to Nvidia GPUs with the GPUDirect protocol . There are much greater storage performance variations between suppliers with GPU Direct than with MLPerf Storage. It would be good to have missing MLPerf Storage performance data for GPUDirect-supporting storage suppliers such as IBM (ESS3500), Huawei (A310), NetApp (A800, EF600) and VAST Data (Ceres). Then we could compare how suppliers look against both benchmarks. MLPerf Inference v3.1 edge inference results can be inspected here and datacenter inference results here . Bootnote The Standard Performance Evaluation Corporation (SPEC) is developing its own machine learning benchmark. It has an ML Committee and its first benchmark, SPEC ML, will measure the end-to-end performance of a system handling ML training and inference tasks. The SPEC ML benchmark will better represent industry practices by including major parts of the end-to-end ML/DL pipeline, including data prep and training/inference. This vendor-neutral benchmark will enable ML users – such as enterprises and scientific research institutions – to better understand how solutions will perform in real-world environments, enabling them to make better purchasing decisions. TAGS
Weka Frequently Asked Questions (FAQ)
When was Weka founded?
Weka was founded in 2013.
Where is Weka's headquarters?
Weka's headquarters is located at 910 East Hamilton Avenue, Campbell.
What is Weka's latest funding round?
Weka's latest funding round is Series D - II.
How much did Weka raise?
Weka raised a total of $273.2M.
Who are the investors of Weka?
Investors of Weka include Qualcomm Ventures, Hewlett Packard Enterprise, NVIDIA, MoreTech Ventures, Hitachi Ventures and 22 more.
Who are Weka's competitors?
Competitors of Weka include Neon, SandStone, Excelero, InfoWorks, Hedvig and 11 more.
Compare Weka to Competitors
OSNEXUS enables organizations to manage their storage ranging from small sites to hyper-scale deployments across global datacenters with its QuantaStor SDS industry-leading Software Defined Storage platform. The QuantaStor SDS platform provides scale-out file, block, and object storage technology as a single, easy-to-manage solution. QuantaStor SDS, deployed by Global 500 companies worldwide, addresses a broad set of storage use cases including server virtualization, big data, cloud computing, and high-performance applications through scale-out physical and virtual storage appliances.
ProphetStor Data Services is a provider of Software-Defined Storage (SDS) and Data Services, offering federated storage and data services to enable both enterprises and cloud service providers to build an agile, automated, intelligent, and orchestrated storage infrastructure. ProphetStor offers integrated multi-use storage systems while enabling the use of commodity servers, by leveraging existing storage and application investments and taking advantage of the future storage investments involving all-flash and big data environments. Its global patent portfolios create the next wave of SDS technology, bringing intelligence into the data center and cloud infrastructure, and providing orchestration, analytic, prediction, and resource automation capabilities that greatly enhance the efficiency of managing data centers.
Scality specializes in data storage software. It built a storage and data management ecosystem to protect customers' data. It offers products such as Ring which provides hybrid cloud storage and Artesca which enables users to store cloud-native objects. The company serves the media and entertainment industry, the healthcare industry, and more. It was founded in 2009 and is based in San Francisco, California.
SandStone provides enterprise-grade software-defined storage products and services. SandStone's SDS is a general virtual platform built on a server, which uses software to manage and adjust the storage resources of the hardware. Its products include block storage, object storage, unified storage, and hyper integration solutions. It was founded in 2014 and is based in Shenzhen, China.
Storpool offers a service dedicated to cutting storage costs for clients by utilizing spare spaces on servers for storage purposes.
DDN is a private data storage company and provider of intelligent technology and infrastructure solutions for Enterprise At Scale, AI and analytics, HPC, government, and academia customers. Through its DDN and Tintri divisions, the company delivers AI, Data Management software and hardware solutions, and unified analytics frameworks to solve complex business challenges for data-intensive, global organizations. DDN provides its enterprise customers with the most flexible, efficient, and reliable data storage solutions for on-premises and multi-cloud environments at any scale.