Sei sulla pagina 1di 36

Introducing the Scan AI Ecosystem

V1
Contents

Introduction to Scan 3
The Scan AI Ecosystem 4
Education and Guided Proof-of-Concept 5
The NVIDIA Deep Learning Institute 5
Guided Proof-of-Concept 6
Deep Learning Hardware 7
Intel AI Solutions 7
NVIDIA DGX-1 – Innovate Faster 8
NVIDIA DGX-2 – The Fastest GPU ever 10
DGX Station Desktop 12
DGX Software and Frameworks 14
IBM POWER9 Solutions 16
IBM POWER9 Platform 17
Optimised Storage 18
NetApp EF570 All-Flash Array 18
NetApp AFF A-series All-Flash Array 19
Pure Storage FlashBlade 20
Introducing AIRI – Artificial Intelligence Ready Infrastructure 22

scan.co.uk/business Contents 1
Contents

Connectivity and Infrastructure 23


Mellanox Infiniband Switches 23
Introducing Eaton Rack Cabinets 24
Software Solutions 25
H2Oai – The AI Platform for Enterprise 25
Kinetica – Instant Insight 27
MapD - Truly Immersive Analytics 29
Planet AI – Speech and Text Searches 30
Clusterone - Scalable Machine Learning Platform for the AI-first World 31
Data Science and Consultancy 32
Bespoke Project Consultancy – Introducing Cambridge Consultants 33
The Scan Business AI Ecosystem Team 34

scan.co.uk/business Contents 2
Introduction to Scan

Welcome to the Scan AI Ecosystem brochure. These


pages are designed to give you an overview of our
comprehensive AI offering – ultimately born out of
decades of experience building and optimising high-
performance professional PC, workstation and server
solutions. Our systems have won countless awards
including a Royal Warrant and we’ve been recognised by
a large number of partners including Intel, Microsoft and
NVIDIA for our quality, expertise and eye for detail.

Our relationships with these industry leaders, amongst


others, have led us into designing and building systems
aimed squarely at the HPC and AI spaces – all the while
backed by our support and driven by the dedication to
provide the most comprehensive solutions to you – our
customers.

Including a dedicated sales team, pre-sales advisors,


solutions architects and data scientists – Scan Business
has grown into the latest division of Scan – and although
these pages are focused on deep learning and AI, Scan
Business encompasses so much more.

scan.co.uk/business Introduction to Scan 3


The Scan AI Ecosystem

Although a high performance hardware-accelerated server is at the centre of deep learning & AI system performance, the efficiency and time
to results is affected by several other factors too.several other factors too. The output from a GPU or FPGA may be limited by the storage array
feeding the dataset, and the connectivity between the server and the storage appliance. Even if you have optimised hardware, deep learning
training results need to be correctly visualised, interpreted and understood – further acceleration of this phase can be delivered by deploying
software in this training phase. Lastly, the key part to building the most appropriate hardware / software solution relies on in-depth knowledge
of the AI environment and understanding what to deploy and when – skilled data scientists and Linux engineers complete the resource required
to deliver fast and insightful deep learning. This brochure introduces each of these areas in greater depth to provide an understanding of how the
various parts of our AI ecosystem make up the whole solution.

scan.co.uk/business The Scan AI Ecosystem 4


Education and Guided Proof-of-Concept

We’ve learned a lot building our AI ecosystem, and we want


to enable our customers with this knowledge to ensure you
understand the full portfolio of solutions on offer to you, but
more importantly how these can aid your AI research, help
to deliver clarity in your results and make you feel sure that
you have the right product mix and expertise on hand for
your project. The following pages will explain how Scan can
provide the basic AI building blocks for your organisation.

The NVIDIA Deep Learning Institute


In order to ensure we share our learnings and experience
in deep learning and AI, we have teamed up with the
NVIDIA Deep Learning Institute and Manchester Science
Partnerships to deliver a full range of courses aimed at
further AI knowledge in business.
Our portfolio includes a range of ‘fundamentals’ courses to
provide a solid understanding of the basics of deep learning
data, visualisation, neural networks and frameworks:

• Deep Learning Demystified


• Fundamentals of Deep Learning
• Fundamentals of Accelerated Computing

scan.co.uk/business Education and Guided Proof-of-Concept 5


These are then followed by industry specific in-depth courses designed to focus on specific areas of development:

• Deep Learning for Autonomous Vehicles


• Deep Learning for Healthcare
• Deep Learning for Intelligent Video Analytics
• Deep Learning for Digital Content Creation
• Deep Learning for Finance

The knowledge gained on these courses will help drive interaction during a proof-of-concept trial and help inform decisions required in the set-up
of your own optimised deep learning or AI ecosystem.

Guided Proof-of-Concept
Deep Learning solutions such as the DGX range of deep learning appliances unlock new possibilities thanks to its unparalleled processing density.
We want you to be sure these solutions are right for you, so we provide the ability to try your own data on one of our deep learning servers as a
Proof of Concept. Hosted in a secure datacentre, we will provide you with remote access to the DGX system of your choice so you can evaluate
and benchmark it.

This service is fully supported by our team of data scientists and Linux engineers to ensure you get the best out of your trial including access to
our software ecosystem embedded on our cloud platform.

scan.co.uk/business Education and Guided Proof-of-Concept 6


Deep Learning Hardware

Every deep learning and AI project is different, and here at Scan we recognise that one size of hardware doesn’t fit all. Depending on the stage
of development you’re at, the applications being used and the time frames results are required in. With this in mind we offer solutions to address
every requirement.

Intel AI Solutions Unleash Full Potential

Intel’s comprehensive, flexible, and Experiences


performance-optimised AI portfolio of products
for machine and deep learning covers the entire
spectrum from hardware platforms to end
user applications. Depending on application Tools Intel Deep Learning Intel Computer Vision SDKM ovidius Neural Saffron
Deployment Toolkit Compute Stick Technology
and dataset, the team behind the Scan AI
Ecosystem are working to optimise Intel Xeon Frameworks Apache mxnet theano Microsoft Torch Caffe
based training systems and FPGA-based Spark CNTK
inferencing systems – both of which will be
available to test shortly within our guided proof- Libraries Intel Dist Intel DAAL Intel Intel MKL Intel Movidius Associative
Nervana MLSL MvTensor Memory
of-concept environment. Graph Library Base

Hardware Memory Networking Visual


& Storage Intelligence

scan.co.uk/business Deep Learning Hardware 7


DGX Station Desktop
The DGX Station is a complementary product to the DGX-1
server, enabling you to develop deep neural networks in a
low latency environment before scaling them out to a data
centre. Unlike earlier desk side deep learning workstations,
the DGX Station is specifically designed for the office, with
special low noise components and a convenient tower
form factor.

Like the DGX-1 server, the DGX Station is based on the


latest NVIDIA Tesla V100 GPUs. These are based on
the innovative Volta architecture that combines the
performance of thousands of traditional CUDA cores with
hundreds of the new Tensor cores that are specifically
optimised for deep learning. With four Tesla V100 cards,
the DGX Station has 20,480 CUDA cores and 2,560
Tensor cores, providing a dramatic performance increase
over previous generation deep learning workstations.

1. NVIDIA Tesla V100

2. Intel Xeon CPU

3. Next Generation NVIDIA NVLink

4. Three Display Ports

5. Water Cooling

6. Half the Power Consumption

scan.co.uk/business Deep Learning Hardware 8


DGX Station Specification

scan.co.uk/business Deep Learning Hardware 9


The NVIDIA DGX-1 – Innovate Faster 1. Network Interconnect
4X InfiniBand 100 Gbps EDR 2X 10 GbE
The NVIDIA DGX-1 deep learning system comprises a
combination of hardware and software that delivers faster 2. GPUs
8X NVIDIA Tesla®V100 16 Gf 40,960
and more accurate training of neural networks. Purpose
Total NVIDIA CUDA® 5,120 Tensor
built for deep learning and AI analytics, the DGX-1 delivers
Cores
performance equivalent to 250 conventional CPU-only
servers.
3. GPU Interconnect
NVIDIA NVLink™
Hybrid Cube Mesh

4. System Memory
512 GB DDR4 LRDIMM

5. CPUs
2X 20-Core Intel® Xeon® E5-2698 v4
2.2 GHz

6. Streaming Cache
The second generation of the DGX-1 is powered by eight 4X 1.92 TB SSDs RAID 0
NVIDIA Tesla V100 GPU accelerators which are based
on the new Volta architecture. These cutting-edge GPUs 7. Power
combine both CUDA cores (5120) and the latest Tensor 4X 1600 W PSUs
Cores (640) plus 32GB of RAM and are specifically [3500 W TOP)
designed for deep learning delivering a massive 5x speed
up compared to the first-generation Pascal-based DGX-1. 8. Cooling
Efficient Front-to-Back Airflow

scan.co.uk/business Deep Learning Hardware 10


DGX-1 Specification

scan.co.uk/business Deep Learning Hardware 11


NVIDIA DGX-2 – The Fastest GPU ever
Experience new levels of AI speed and scale with NVIDIA
DGX-2, the first 2 petaFLOPS system that combines
16 fully interconnected GPUs for 10x the deep learning
performance. It’s powered by NVIDIA DGX software and
a scalable architecture built on NVIDIA NVSwitch, so
you can take on the world’s most complex AI challenges.
The DGX-2 is only available from NVIDIA Elite Solution
providers.

The DGX-2 is powered by sixteen NVIDIA Tesla V100


GPU accelerators, which are based on the new Volta
architecture. These cutting-edge GPUs combine both
CUDA cores (5120) and the latest Tensor Cores (640)
plus 32GB of RAM and are specifically designed for deep
learning delivering a massive 10x speed up compared to
the DGX-1. Leading edge research demands the freedom
to leverage model parallelism and requires never-before- 1. NVIDIA Tesla V100 32GB 4. Eight EDR Infiniband/100 GigE
seen levels of inter-GPU bandwidth. NVIDIA has created 4X InfiniBand 100 Gbps EDR 2X 10 GbE 1600 Gb/sec Total Bi-directional Bandwidth
NVSwitch to address this need. Like the evolution from
dial-up to ultra-high speed broadband, NVSwitch delivers 2. Two GPU Boards 5. PCIe Switch Complex
a networking fabric for the future, today. 8 V100 32GB GPUs per board 6 NVSwitches per
board 512GB Total HBM2 Memory interconnected 6. Two Intel Xeon Platinum CPUs
by Plane Card
7. 1.5 TB System Memory
3. Twelve NVSwitches
2.4 TB/sec bi-section bandwidth 8. 30 TB NVME SSDs
Internal Storage

9. Dual 10/25 GigE

scan.co.uk/business Deep Learning Hardware 12


DGX-2 Specification

scan.co.uk/business Deep Learning Hardware 13


DGX Software and Frameworks
While it’s entirely possible to write / code your own Deep Learning training and deployment system from scratch, it’s worth considering some of
the other options out there; one option is to use NVIDIA’s cuDNN library in your own code. cuDNN is freely available to registered developers, and
is optimised to an extreme level to run as fast as possible on NVIDIA GPUs by the people who make those GPUs. Another option is to use one of a
number of Deep Learning frameworks.

Deep Learning frameworks are systems for the training and deployment of Deep Learning networks which provide the flexibility to design and
adapt the network to your specific task, frameworks for Deep Learning allow you to hit the ground running, to prototype and test ideas and
applications without the considerable overhead of writing significant amounts of your own code. All the major frameworks use cuDNN so you can
rely on optimised code, and each one has a community of developers and users who can help you not only get the most from the framework you
chose but also guide you in modifying the framework to provide new features you may require for your application.

Software
The DGX family is much more than a range of GPU appliances - they
are deep learning solutions comprising of a finely tuned combination of
hardware and software. Running a GPU-optimised version of Ubuntu
Server Linux, the software stack comprises drivers, the NVDocker
container tool, deep learning SDK, NVIDIA Cloud Management Service
plus NVIDIA DIGITS which is used to run deep learning frameworks such
as Caffe, Torch, TensorFlow and many more.

The Operating System (Ubuntu Server Linux) is optimised to take


advantage of the hardware and software features of the system and
CUDA 8, especially with respect to memory management and hardware
communications.

The deep learning frameworks provided with the system are especially
optimised to take advantage of the NVlink communication links among
other enhancements, in order to optimize multi-GPU communication in
the system.

scan.co.uk/business Deep Learning Hardware 14


Deep Learning Frameworks

Framework From

Berkley Caffe provides an easy and accessible way to define and train deep neural networks via a high-level scripting language (Google’s Protocol Buffer) describing the network.
Vision and
Learning Extending your application beyond the pre-defined Caffe modules is fairly straight forward but will require programming your own Caffe modules. Caffe is a powerful
Centre command line driven framework.

DIGITS isn’t really a framework in its own right but rather provides a powerful graphical front end to both Caffe and Torch, simplifying interaction, setup, and providing
useful visualisations of what’s going on in your deep neural network.
NVIDIA
DIGITS also helps optimise the use of multiple NVIDIA GPUs.

In contrast to Caffe’s high-level script, Torch throws you in at a deeper level requiring you to program your Deep Learning model while providing a powerful and useful
level of description.
Facebook
This can make Torch appear more flexible than Caffe as it is clearer how the entire learning process operates, as you have had to define each step. Torch is also a
command line and programming driven framework.

TensorFlow was developed as part of the Google Brain project as a framework to work with multi-dimensional arrays (tensors) utilising data flow graphs to solve machine
learning and deep learning problems.
Google
TensorFlow aims to provide a more professional tool for developing and managing Deep Learning, it also provides its own graphical interface simplifying interaction,
setup, and providing useful visualisations of what’s going on in your deep neural network.

CNTK (Cognitive toolkit) according to Microsoft is “a unified deep-learning toolkit that describes neural networks as a series of computational steps via
a directed graph”. It is another alternative for the other deep learning frameworks mentioned. However a key distinguishing factor is that the framework supports
Microsoft
parallelisation across both multiple machines and multiple GPUs (without regard to where the GPUs are located). The goal of this framework is to provide efficiency,
performance (fast training and productisation) and flexibility (application to speech, vision, text).

scan.co.uk/business Deep Learning Hardware 15


The IBM POWER9 Solutions
The IBM POWER9 processor - built from the ground-up for data intensive workloads, POWER9 is the only processor with state-of-the-art I/O
subsystem technology, including next generation NVIDIA NVLink, PCIe Gen4, and OpenCAPI. Combined with up to six class-leading NVIDIA GPU
accelerator cards to deliver an agile AI experience – and integrated into the Scan AI ecosystem to offer comprehensive data scientist, engineer
and configuration support.

Accelerated Computing
5x Higher Energy Efficiency

AC922

MTM8 335-GTH | 8335-GTX

System Packaging 2U

Processor Socket2 S

# of cores Up to 44 cores

Number of GPUs 4 or 6 NVIDIA Tesla GPU processors

Memory DIMM Slots 16

Memory-Max 1TB

HDD/SDDT wo SFF 2.5” SATA Drives, Max 4TB HDD, Max 7.68TB SSD

PCIe G4 Slot 4 Slots

scan.co.uk/business Deep Learning Hardware 16


The IBM POWER9 Platform
PowerAI makes deep learning, machine learning, and AI more accessible and more performant. By combining this software platform for deep
learning with IBM Power Systems, enterprises can rapidly deploy a fully optimized and supported platform for AI with blazing performance.

scan.co.uk/business Deep Learning Hardware 17


Optimised Storage

Through extensive testing and benchmarking within our own proof-of concept environment we have optimised a number of flash-based data
management solutions that work alongside our AI server platforms to provide the required datasets at a fast enough pace to deliver training results
that address every budget and performance requirement.

NetApp EF570 All-Flash Array


Capable of 100Gb InfiniBand, or 100Gb NVMe over InfiniBand, the NetApp EF570 makes an ideal partner for a single or multiple (partitioned)
NVIDIA DGX-1 appliances where speed of the storage platform is key. The NetApp all-flash EF570 provides an industry-leading price /
performance in an enterprise-grade 2U system. With support for up to 367TB of flash capacity in a single modular 2U building block, the ultra-
dense EF570 enables you to easily address ever-evolving AI & deep learning world. Additionally, the EF570 provides investment protection to
meet future demands without forklift upgrades through the ability to independently scale to 1.8PB of raw flash capacity.

scan.co.uk/business Optimised Storage 18


NetApp AFF A-series All-Flash Array
Designed specifically for flash, the AFF A-Series all-flash systems deliver industry-leading performance, capacity density, scalability, security,
and network connectivity in dense form factors – ideal for partnering with a DGX cluster environment. With the addition of a new entry-level
system, the new AFF A-Series family extends enterprise-grade flash to midsize businesses and to fit any budget. At up to 7M IOPS per cluster
with sub-millisecond latency, they are the fastest all-flash arrays built on a true unified scale-out architecture. Best suited to a DGX cluster
environment.

Model AFF A700s AFF A700 AFF A300 AFF A200


NAS Scale-out 2-24 nodes (12 HA pairs) 2-24 nodes (12 HA pairs) 2-24 nodes (12 HA pairs) 2-8 nodes (4 HA pairs)

Maximum SSD 2592 5760 4608 576

Maximum Raw Capacity 39.0 PB/35.2PiB 88.1PB/78.3PiB 70.5PB/62.6PiB 8.8PB/7.8PiB

Effective Capacity 155.5PB/138.1PiB 356.3PB/316.4PiB 285.3PB/253.1PiB 35.7PB/30.8PiB

scan.co.uk/business Optimised Storage 19


Pure Storage FlashBlade
Modern workloads, like deep learning and AI, are massively parallel – and designed to run on parallel computers with parallel storage for the
highest performance. FlashBlade is architected from the ground-up for modern data analytics – delivering powerful cloud-era flash that’s fast, big,
and simple.

The Elements of Flashblade

Blade Purity Fabric


Modern workloads, like deep learning The heart of FlashBlade, Includes a built in 40Gb Ethernet
and AI, are massively parallel – and implementing its scale-out fabric providing a total network
designed to run on parallel computers storage capabilities, services and bandwidth of 320Gb/s for the chassis.
with parallel storage for the highest management.
performance.

scan.co.uk/business Optimised Storage 20


Capacity 8TB Blade 17TB Blade 52TB Blade

7 Blades 98 TBs Usable 190 TBs Usable 591 TBs Usable

15 Blades 267 TBs Usable 525 TBs Usable 1607 TBs Usable

scan.co.uk/business Optimised Storage 21


Introducing AIRI – Artificial
Intelligence Ready Infrastructure
Engineered as a fully integrated software and hardware
solution by Pure Storage and NVIDIA, AIRI shatters
infrastructure complexities that hold organizations
back from deploying AI-at-scale. AIRI is powered by
FlashBlade, and paired with four DGX-1 supercomputers,
delivering four petaFLOPS of performance. These
systems are interconnected with Arista 100GbE switches
and the whole system is supported by the NVIDIA GPU
Cloud deep learning stack and Pure Storage AIRI Scaling
Toolkit, enabling data scientists to jumpstart their AI
initiatives in hours, not weeks or months.

scan.co.uk/business Optimised Storage 22


Connectivity and Infrastructure

You have the high performance server and optimised all-flash storage array – but how to be sure you create an infrastructure that doesn’t
introduce any bottlenecks when connecting the components together. Once again our experienced team has tested various configurations so
we’re able to advise the best options for your needs.

Mellanox Infiniband Switches


NVIDIA’s DGX range of deep learning appliances are designed to deliver maximum performance and as such the connectivity architecture within
is based on the four Mellanox ConnectX®-4 EDR 100Gb/s InfiniBand HCAs. DGX appliances leverage multiple EDR InfiniBand ports to eliminate
performance bottlenecks and to provide maximum data throughput at ultra-low latencies, so it makes sense that when constructing your AI
ecosystem Mellanox Infiniband switches are the obvious choice to connect your chosen optimised storage platform.

Switch-IB, the seventh generation of high performance switching from Mellanox, is renowned for industry leading bandwidth, low-latency,
and scalability. The devices supports up to 36 EDR 100Gb/s InfiniBand ports, and all ports comply with the InfiniBand specification for auto-
negotiation SDR to EDR.

scan.co.uk/business Connectivity and Infrastructure 23


Introducing Eaton Rack Cabinets
Available in a range of sizes & heights, and designed for
application-specific configurations, the new Eaton RS
Enclosure provides an easy-to-configure rack solution
for IT equipment storage – perfect for your AI ecosystem.
Features include pathways for cable integration at the
top and bottom of the enclosure to accommodate large
connectors and cable loops, tool-less options to configure
RS components, including mounting rails, PDUs and PDU
brackets, doors, top panels, side panels, flexible cable
management solutions and airflow management.

scan.co.uk/business Connectivity and Infrastructure 24


Software Solutions

Every deep learning and AI project is different, and here at Scan we recognise that one size of hardware doesn’t fit all. Depending on the stage
of development you’re at, the applications being used and the time frames results are required in. With this in mind we offer solutions to address
each phase of AI development.

H2Oai – The AI Platform for Enterprise


Combine the power of highly advanced algorithms, the freedom of open source, and truly scalable in-memory processing for big data on one or
many nodes.

Fast & Accurate Scalable Easy to Use


The responsiveness of in-memory Fine-Grain distributed processing on big Get started quickly using H2O’s
processing and the ability to run fast data at speeds up to 100x faster is done intuitive web-based Flow GUI or familiar
serialisation between nodes and clusters with fine-grain parallelism, which enables programming environments. Deploy
are combined—so you can support the size optimal efficiency, without introducing POJOs and MOJOs, score new data for
requirements of your large data sets. degradation in computational accuracy. accurate predictions in any environment.

scan.co.uk/business Software Solutions 25


Driverless AI – AI to do AI
Driverless AI by H2O.ai is a machine learning platform that automates many of the most difficult data science and machine learning workflows,
such as feature engineering, model validation, model tuning, model selection and model deployment.

It aims to achieve the highest predictive accuracy, comparable to expert data scientists, but in much shorter time thanks to end-to-end
automation. Driverless AI also offers automatic visualisations and machine learning interpretability. Especially in regulated industries, model
transparency and explanation are just as important as predictive performance.

The community version of Driverless AI is a fully functional version with a 30-day trial on DGX systems and supported NGC platforms. During
your proof-of-concept trial you can achieve up to 40x speedups on GPU-accelerated algorithms vs. on CPUs. Driverless AI provides speed,
accuracy and interpretability. Driverless AI empowers you to do automatic feature engineering and interpret debugging with reason codes in plain
English.

scan.co.uk/business Software Solutions 26


Kinetica – Instant Insight
Kinetica’s distributed, in-memory database on NVIDIA DGX-1 and NVIDIA GPUs provides real-time analytics on data in motion and at rest
providing 10-100x faster performance at 1/10 of the cost of traditional systems. NVIDIA and Kinetica together deliver unmatched performance,
predictable scalability across multiple high-density nodes, and seamless integration with industry-standard connectors to data sources and
applications. Kinetica’s User Defined Functions (UDFs) further deliver the first converged AI and BI workloads accelerated by NVIDIA GPUs.

Unparalleled Speed Streaming Data Analysis Visual Foresight Streamlined Machine


Learning
Dramatically accelerate analysis Ingest and process data at the Instantly translate temporal,
with advanced parallel computing same time to get instant results geospatial, and streaming data Train models faster and predict
into visuals that reveal patterns more accurately with integrated
• In-memory GPU database • Run SQL queries on streaming and opportunities machine learning workflows
with advanced GPU abstraction and geospatial data
technology • In-memory, distributed image • Seamlessly integrate pre-
• Continuously collect, analyse, processing & rendering existing code with highly scalable
• SQL queries to process and and integrate streaming data with solution
analyze billions of rows in historical data • Interactive visual discovery
microseconds optimised for temporal & • Simultaneously process &
• All-in-one solution to ingest, geospatial analysis manage BI and AI workloads
• Simpler data preparation compute, visualise, and egress
through distributed ingestion and data • Integrated visual dashboard • Integrate workflows with open
extraction without the need to source machine learning libraries
index

scan.co.uk/business Software Solutions 27


One of the core challenges with geospatial analytics is moving data from the database layer to the visualisation layer. Serialising and moving
millions to billions of objects from one technology to another takes time. Kinetica is able to short-circuit this by keeping the data within its
database, executing complex geospatial filters and advanced analytics, and rendering the geospatial data on the fly through its internal geospatial
web server. With GPU architecture, it is possible to filter and visualise large, complex geospatial vector data at high speed, and without the
challenges of moving data from a database to a separate geospatial server layer.

scan.co.uk/business Software Solutions 28


MapD - Truly Immersive Analytics
GPU-accelerated analytics applications are now available in the NVIDIA DGX container registry and NVIDIA GPU Cloud (NGC). These
applications, including MapD, provide customers the ability to abstract insights in milliseconds, build models with transparency and accuracy,
and eliminate any integration complexity. They are tested and deployed on DGX systems and supported NGC platforms and are available to use
immediately as part of your Scan proof-of-concept trial.

MapD is a GPU-accelerated platform with an open-source SQL engine called MapD Core and an integrated visualisation system called MapD
Immerse. Open source MapD Core is now containerised on DGX systems and in the NGC. Customers registering for a Scan proof-of-concept trial
on the DGX platforms can use the software as part of their experience, to understand how MapD could accelerate their insights. Data scientists
using MapD experience unparalleled analytic speed, constant innovation from the open source community, and interactive visual exploration of
the data used to build machine learning models.

MapD Immerse

Streaming Data Rendering Engine SQL Machine Learning

S3, Data Lake,


HDFS MapD Core SQL Engine Data Science

3rd Party Viz


Data Warehouse Custom Apps

Input GPU Acceleration Output

scan.co.uk/business Software Solutions 29


Planet AI – Speech and Text Searches
Planet AI is a team of scientists and engineers with deep roots in Artificial Intelligence, Machine Learning and Cognitive Computing undertaking its
own ambitious research projects towards Deep Universal Sequence Understanding. These skills are accomplished with many experiences in the
domains of image and signal processing and has resulted in winning competitions over three years in the field of handwriting recognition, keyword
spotting and document analysis.

Text Recognition
Archiving documents and recorded calls is one thing, but the real power comes in being to effectively and
efficiently search those archives and extract the key information required. ArgusSearch software from
Planet AI allows ‘search-engine-like’ display of information found in historical and complex handwritten
or hand-printed documents. Unlike conventional ICR that provides either a transcribed word or a ‘not read’
result, ArgusSearch provides ‘quasi’ matches for a particular search term based on the equivalence of the
search term and a particular instance of the term within a document.

Speech Recognition
Understanding speech content and making it searchable rather just having a huge depository of digital files
is key. ArgusSpeech software from Planet AI can detect phrases, written or spoken in millions of stored
audio recordings and video recordings returning results where the phrase is spoken in the recordings. No
more manual listening tohuge numbers of voice recordings to find specific keywords and then only listen to
those recordings where the keywords exists. Alternatively let the software do the hard work of continuously
trawling through recordings to search a big list of keywords.

scan.co.uk/business Software Solutions 30


Clusterone - Scalable Machine Learning Platform for the AI-first World
Clusterone enables deep learning teams to be efficient through faster model training and automated workflows. It provides distributed training
management, advanced cloud resources orchestration and productivity tools. Deep learning scientists and engineers can be productive from Day
1, and spend time on modelling, not setup and DevOps. Clusterone runs on top of any infrastructure: in the cloud, on-premise, or both, and has
been deployed and tested by Scan on our DGX cluster enabling us to offer best practice in advancing your deep learning.

Clusterone can speed up training times using clustered hardware by using Data Parallelism – where each machine trains on a separate set of
training data, leading to significant increases in training speed. The most common approach is to use one machine to store the model parameters.
This parameter server sends parameters of the model to multiple worker machines, who each run the training procedure on a small batch of data
and then return the updates to the model parameters back to the parameter server.

scan.co.uk/business Software Solutions 31


Data Science and Consultancy

We’ve worked hard to understand what delivery of a complete AI project requires – and often the key element is the expertise to get all
component parts working seamlessly together. System power is nothing without fast data ingress; a short time to results has no bearing if a lack
of visualisation offers no insight – our team of data scientists, engineers and architects are there to help smooth all these technology interfaces to
guarantee the best experience – whatever the scale

In addition to the infrastructure we have a full team of expert consultants including data scientists, Linux engineers and hardware architects to
supported your AI development whether it be a proof-of-concept trial, project planning or full deployment. These specialist teams are available in
half-day increments to hand-hold your deep learning projects, or simply to compliment your in-house experts as and when you require.

Data Scientist Infrastructure Engineer Security Consultancy Project Consultancy


Consultancy Consultancy
For help with the security and Ongoing support for longer term
For help in hardware regulatory compliance aspects of projects that may encompass
For help and advice with all
configuration or software (Linux) your deep learning project. skills sets from all of the above
frameworks, algorithms and
set up of your deep learning areas.
interpretation of your deep
learning and AI results. environment.

scan.co.uk/business Data Science and Consultancy 32


Bespoke Project Consultancy – Introducing Cambridge Consultants
Although our expertise and specialist deep learning team at Scan is always expanding, we recognise when a project is large or complex enough to
require a dedicated AI think tank – additional resource in the form of Cambridge Consultants.

Machine learning is at the core of a new wave of artificial intelligence applications limited only by our imagination. New algorithmic approaches,
recent jumps in processing power and large training data sets generated by internet users mean that, for the first time, machines can learn to solve
useful problems without explicit programming.

Like Scan’s own proof-of-concept platform, their platform runs on high-performance computing based around NVIDIA’s DGX-1 Deep Learning
Supercomputer and other GPU- and FPGA-accelerated servers, providing petaflop-scale compute on-site. This links to petabyte-scale local storage,
project-specific clouds and our continuous integration systems. When the organically grown deep learning is ready, Cambridge Consultants can
export models easily into customers’ own compute facilities or the cloud.

scan.co.uk/business Data Science and Consultancy 33


The Scan Business AI Ecosystem Team

Head of AI Deep Learning & Principal Solutions Principal


Business Unit AI Sales Specialist Architect Data Scientist
Sunil Mistry Dave Green Eyal Lemberger Dr Soumik Ghosh

Scan Computers is proud to Scan Computers is proud to


announce that we have been announce that we have been
awarded a Cyber Essentials accredited as an Investor in People.
certificate.

scan.co.uk/business Data Science and Consultancy 34


To learn more contact business@scan.co.uk
scan.co.uk/business • 01204 47 47 47

Potrebbero piacerti anche