BITMAIN SM5 (SOPHON SM5) is an AI computing module with super computing power. It is positioning the edge computing scenes with high performance requirements and has AI analysis capabilities of over 16 channels FHD video .
BITMAIN's SOPHON SM5 is equipped with the third-generation TPU chip BM1684 independently developed by BITMAIN.It has computing power of up to 17.6TOPS at INT8 and can process over 16 FHD videos simultaneously. Its size is like a credit card, and it has rich IO interfaces, which make it easy to be integrated into various edge or embedded devices.The tool chain is complete and easy to use, and the algorithm migration cost is small.
Chip BM1684
17.6TOPS
38 channel video hardware decoding
17.6 TOPS INT8 computing power, up to 35.2 TOPS with Winograd convolution acceleration, far superior than similar products in the industry. The typical power consumption of 16-channel video stream analysis is lower than 16W
Support up to 38 channels [email protected] 1080P decoding, support H264/H265 format, can realize over 16 channels FHD video stream face detection analysis or video structuring
Support PCIE slave mode and SOC host mode, support FP32 high precision and INT8 low precision
Support the mainstream framework of AI industry Caffe, tensorflow, Pytorch, paddle, Mxnet
It is applied to intelligent public security, intelligent park, intelligent retail, intelligent power, industrial robot UAV and other visual computing AI scenes.
BMNNSDK (BITMAIN Neural Network SDK) one-stop toolkit provides a series of software tools such as the underlying driver environment, compiler, inference deployment tool and so on. Easy to use and convenient, covering the model optimization, efficient runtime support and other capabilities required for the neural network inference stage, providing easy-to-use and efficient full-stack solutions for deep learning application development and deployment. BMNNSDK minimizes the development cycle and cost of algorithms and software. Users can quickly deploy deep learning algorithms on various AI hardware products of Fortune Group to facilitate intelligent applications.
AI chip
1 BM1684
AI computing power
FP32(FLOPS)
2.2TFLOPS
INT8(TOPS) Winograd OFF
17.6TOPS
INT8(TOPS) Winograd ON
35.2TOPS
Memory configuration
Standard configuration
12GB
CPU capacity
CPU (SOC host mode)
8-core ARM A53, 2.3GHz main frequency
High speed data interface
PCIE EP interface
PCIE 3.0, X4
(connector interface)
PCIE RC interface
PCIE 3.0, X4
Ethernet ports
Dual Gigabit Ethernet ports
Video decoding format
H.264 and H.265
Maximum decoding resolution
Support 4K, 8K (semi-real time)
Picture decoding and encoding performance
480 PCS/sec @1080p
Low speed data interface
RS485 / RS232 / GPIO / SDIO / PWM / I2C etc.
Connector
144-pin connector
Power consumption
Typical power consumption <20W
Maximum power consumption 25W
Heat dissipation mode
SM5-P includes passive heatsink
SM5-A includes active cooling fan
L x W x H
87 x 65 x 8mm without heatsink
92 x 70 x 20.1mm (SM5-P includes passive heatsink)
92 x 70 x 20.1mm (SM5-A includes active cooling fan)