QNAP Mustang-F100

AED 7,563.00 (Inc. VAT)
Brand: QNAP       SKU: Mustang-F100-A10-R10       GTIN: 842936100887      

Pre-Order
 Sold out but we have more in transit and may have a close replacement in stock. Place the order and we promise to get back to you with the exact delivery time or a similar product

QNAP Mustang-F100

AED 7,563.00 (Inc. VAT)
Brand: QNAP       SKU: Mustang-F100-A10-R10       GTIN: 842936100887      

Pre-Order
 Sold out but we have more in transit and may have a close replacement in stock. Place the order and we promise to get back to you with the exact delivery time or a similar product
Free Delivery within UAE Customer service to update Delivery ETA

More Information about QNAP Mustang-F100 Mustang-F100-A10-R10

Logistics, Bulk pricing , Stock and Partnership Info

QNAP One of the Largest Subdistributor and Reseller partner in UAE

We are one of the largest QNAP subdistributor offering complete Storage solution with QNAP using hard drives from Seagate , WD or Toshiba. Every QNAP device you buy from us with hard drives will be pre-configured for free. Other value-added services for QNAP devices such as warranty extension for new QNAP devices (QNAP Warranty Extension), warranty extension for existing QNAP devices (PNP Warranty Extension), extended replacement service with SLA ( NBD Hardware Replacement), etc. are offered in UAE at an additional cost.
Services such as NBD replacement of a failed QNAP device, warranty extensions for existing QNAP devices are available exclusively from us at an additional cost due to our service partner agreement with QNAP. Our Engineers are trained by QNAP and we also arrange workshops and techincal trainings for QNAP.

Stock availability of QNAP Mustang-F100 Mustang-F100-A10-R10 Dubai UAE

We have a large inventory (probably one of the largest in the Middle East) of QNAP as we are a sub-distributor. If the stock is not available, we can give you a firm delivery date. Contact us and we will let you know the exact availability.

Price Matching and Lowest Bulk Price for QNAP Mustang-F100 Mustang-F100-A10-R10

The price listed on the website is the retail price intended for end users, but If you are an end user looking for a better price you can contact us. Resellers can request for a special reseller price.

Free Shipping for QNAP Mustang-F100 Mustang-F100-A10-R10 in UAE

We Ship QNAP Mustang-F100 Mustang-F100-A10-R10 to Dubai, Abu Dhabi and Rest of the UAE free of cost. We have Export freight rates for this product also to Middle East including Saudi Arabia, Kuwait, Qatar, Bahrain and Oman. For African countries, our organized carriers tend to be expensive and we ask that you arrange shipping yourself. VAT Exceptions for export orders from the UAE are available if shipped from Jebel Ali or if your carrier issues proper export documentation

Information about QNAP Mustang-F100 from Vendor Site

  • PCIe-based accelerator card
  • Boost your NAS/PC Computing power
  • Half-height, half-length, single-slot compact size

--
overview

Mustang-F100

Intel Vision Accelerator Design with Intel Arria 10 FPGA

As QNAP NAS evolves to support a wider range of applications (including surveillance, virtualization, and AI) you not only need more storage space on your NAS, but also require the NAS to have greater power to optimize targeted workloads. The Mustang-F100 is a PCIe-based accelerator card using the programmable Intel Arria 10 FPGA that provides the performance and versatility of FPGA acceleration. It can be installed in a PC or compatible QNAP NAS to boost performance as a perfect choice for AI deep learning inference workloads.

  • Half-height, half-length, double-slot.
  • Power-efficiency, low-latency.
  • Supported OpenVINO toolkit, AI edge computing ready device.
  • FPGAs can be optimized for different deep learning tasks.
  • Intel FPGAs supports multiple float-points and inference workloads.

Available Models : Mustang-F100-A10-R10

PCIe FPGA Highest Performance Accelerator Card with Arria 10 1150GX support DDR4 2400Hz 8GB, PCIe Gen3 x8 interface


OpenVINO toolkit

OpenVINO toolkit is based on convolutional neural networks (CNN), the toolkit extends workloads across Intel hardware and maximizes performance. It can optimize pre-trained deep learning model such as Caffe, MXNET, Tensorflow into IR binary file then execute the inference engine across Intel-hardware heterogeneously such as CPU, GPU, Intel Movidius Neural Compute Stick, and FPGA.


Get deep learning acceleration on Intel-based Server/PC

You can insert the Mustang-F100 into a PC/workstation running Linux (Ubuntu) to acquire computational acceleration for optimal application performance such as deep learning inference, video streaming, and data center. As an ideal acceleration solution for real-time AI inference, the Mustang-F100 can also work with Intel OpenVINO toolkit to optimize inference workloads for image classification and computer vision.

  • Operating Systems
    Ubuntu 16.04.3 LTS 64-bit, CentOS 7.4 64-bit, Windows 10 (More OS are coming soon)
  • OpenVINO toolkit
    • IntelDeep Learning Deployment Toolkit
      • - Model Optimizer
      • - Inference Engine
    • Optimized computer vision libraries
    • IntelMedia SDK
      *OpenCL graphics drivers and runtimes.
    • Current Supported Topologies: AlexNet, GoogleNet, Tiny Yolo, LeNet, SqueezeNet, VGG16, ResNet (more variants are coming soon)
    • IntelFPGA Deep Learning Acceleration Suite
  • High flexibility, Mustang-F100-A10 develop on OpenVINO toolkit structure which allows trained data such as Caffe, TensorFlow, and MXNet to execute on it after convert to optimized IR.

QNAP NAS as an Inference Server

OpenVINO toolkit extends workloads across Intel hardware (including accelerators) and maximizes performance. When used with QNAPs OpenVINO Workflow Consolidation Tool, the Intel-based QNAP NAS presents an ideal Inference Server that assists organizations in quickly building an inference system. Providing a model optimizer and inference engine, the OpenVINO toolkit is easy to use and flexible for high-performance, low-latency computer vision that improves deep learning inference. AI developers can deploy trained models on a QNAP NAS for inference, and install the Mustang-F100 to achieve optimal performance for running inference.

Note:

1. QTS 4.4.0 (or later) and OWCT v1.1.0 are required for the QNAP NAS.

2. To use FPGA card computing on the QNAP NAS, the VM pass-through function will be disabled. To avoid potential data loss, make sure that all ongoing NAS tasks are finished before restarting.


Easy-to-manage Inference Engine with QNAP OWCT

Upload a video file

Download inference result

Specifications

Check Compatible NAS Models

28-Bay TS-2888X
24-Bay TVS-2472XU-RP
16-Bay TVS-1672XU-RP
12-Bay TVS-1272XU-RP
8-Bay TVS-872XU, TVS-872XU-RP

Mustang-F100-A10-R10

Main FPGA Intel Arria 10 GX1150 FPGA
Operating Systems PC: Ubuntu 16.04.3 LTS 64-bit, CentOS 7.4 64-bit, Windows 10 (More OS are coming soon)
NAS: QTS (Installing Mustang Card User Driver in the QTS App Center is required.)
Voltage Regulator and Power Supply Intel Enpirion Power Solutions
Memory 8G on board DDR4
Dataplane Interface PCI Express x8
Compliant with PCI Express Specification V3.0
Power Consumption (W) <60W
Operating Temperature & Relative Humidity 5C~60C (ambient temperature)5% ~ 90%
Cooling Active fan: (50 x 50 x 10 mm) x 2
Dimensions 169.5 mm x 68.7 mm x 33.7 mm
Power Connector *Preserved PCIe 6-pin 12V external power
Dip Switch/LED indicator Up to 8 cards can be supported with operating systems other than QTS; QNAP TS-2888X NAS supports up to 4 cards. Please assign a card ID number (from 0 to 7) to the Mustang-F100 by using rotary switch manually. The card ID number assigned here will be shown on the LED display of the card after power-up.
FAQ
Downloads

Description, Specifications for QNAP Mustang-F100

  • PCIe-based accelerator card
  • Boost your NAS/PC Computing power
  • Half-height, half-length, single-slot compact size

--
overview

Mustang-F100

Intel Vision Accelerator Design with Intel Arria 10 FPGA

As QNAP NAS evolves to support a wider range of applications (including surveillance, virtualization, and AI) you not only need more storage space on your NAS, but also require the NAS to have greater power to optimize targeted workloads. The Mustang-F100 is a PCIe-based accelerator card using the programmable Intel Arria 10 FPGA that provides the performance and versatility of FPGA acceleration. It can be installed in a PC or compatible QNAP NAS to boost performance as a perfect choice for AI deep learning inference workloads.

  • Half-height, half-length, double-slot.
  • Power-efficiency, low-latency.
  • Supported OpenVINO toolkit, AI edge computing ready device.
  • FPGAs can be optimized for different deep learning tasks.
  • Intel FPGAs supports multiple float-points and inference workloads.

Available Models : Mustang-F100-A10-R10

PCIe FPGA Highest Performance Accelerator Card with Arria 10 1150GX support DDR4 2400Hz 8GB, PCIe Gen3 x8 interface


OpenVINO toolkit

OpenVINO toolkit is based on convolutional neural networks (CNN), the toolkit extends workloads across Intel hardware and maximizes performance. It can optimize pre-trained deep learning model such as Caffe, MXNET, Tensorflow into IR binary file then execute the inference engine across Intel-hardware heterogeneously such as CPU, GPU, Intel Movidius Neural Compute Stick, and FPGA.


Get deep learning acceleration on Intel-based Server/PC

You can insert the Mustang-F100 into a PC/workstation running Linux (Ubuntu) to acquire computational acceleration for optimal application performance such as deep learning inference, video streaming, and data center. As an ideal acceleration solution for real-time AI inference, the Mustang-F100 can also work with Intel OpenVINO toolkit to optimize inference workloads for image classification and computer vision.

  • Operating Systems
    Ubuntu 16.04.3 LTS 64-bit, CentOS 7.4 64-bit, Windows 10 (More OS are coming soon)
  • OpenVINO toolkit
    • IntelDeep Learning Deployment Toolkit
      • - Model Optimizer
      • - Inference Engine
    • Optimized computer vision libraries
    • IntelMedia SDK
      *OpenCL graphics drivers and runtimes.
    • Current Supported Topologies: AlexNet, GoogleNet, Tiny Yolo, LeNet, SqueezeNet, VGG16, ResNet (more variants are coming soon)
    • IntelFPGA Deep Learning Acceleration Suite
  • High flexibility, Mustang-F100-A10 develop on OpenVINO toolkit structure which allows trained data such as Caffe, TensorFlow, and MXNet to execute on it after convert to optimized IR.

QNAP NAS as an Inference Server

OpenVINO toolkit extends workloads across Intel hardware (including accelerators) and maximizes performance. When used with QNAPs OpenVINO Workflow Consolidation Tool, the Intel-based QNAP NAS presents an ideal Inference Server that assists organizations in quickly building an inference system. Providing a model optimizer and inference engine, the OpenVINO toolkit is easy to use and flexible for high-performance, low-latency computer vision that improves deep learning inference. AI developers can deploy trained models on a QNAP NAS for inference, and install the Mustang-F100 to achieve optimal performance for running inference.

Note:

1. QTS 4.4.0 (or later) and OWCT v1.1.0 are required for the QNAP NAS.

2. To use FPGA card computing on the QNAP NAS, the VM pass-through function will be disabled. To avoid potential data loss, make sure that all ongoing NAS tasks are finished before restarting.


Easy-to-manage Inference Engine with QNAP OWCT

Upload a video file

Download inference result

Specifications

Check Compatible NAS Models

28-Bay TS-2888X
24-Bay TVS-2472XU-RP
16-Bay TVS-1672XU-RP
12-Bay TVS-1272XU-RP
8-Bay TVS-872XU, TVS-872XU-RP

Mustang-F100-A10-R10

Main FPGA Intel Arria 10 GX1150 FPGA
Operating Systems PC: Ubuntu 16.04.3 LTS 64-bit, CentOS 7.4 64-bit, Windows 10 (More OS are coming soon)
NAS: QTS (Installing Mustang Card User Driver in the QTS App Center is required.)
Voltage Regulator and Power Supply Intel Enpirion Power Solutions
Memory 8G on board DDR4
Dataplane Interface PCI Express x8
Compliant with PCI Express Specification V3.0
Power Consumption (W) <60W
Operating Temperature & Relative Humidity 5C~60C (ambient temperature)5% ~ 90%
Cooling Active fan: (50 x 50 x 10 mm) x 2
Dimensions 169.5 mm x 68.7 mm x 33.7 mm
Power Connector *Preserved PCIe 6-pin 12V external power
Dip Switch/LED indicator Up to 8 cards can be supported with operating systems other than QTS; QNAP TS-2888X NAS supports up to 4 cards. Please assign a card ID number (from 0 to 7) to the Mustang-F100 by using rotary switch manually. The card ID number assigned here will be shown on the LED display of the card after power-up.
FAQ
Downloads

FREE Delivery in UAE

Including Dubai, Abu Dhabi and Sharjah

NAS Experts

QNAP , Synology Top Sub Distributor and Service Partners in UAE

TOP Integrator

Second time Winner of GEC Awards in Data Storage Category UAE

Honest and Credible

4.8+ Star Google rating. Thousands of satisfied customers in UAE