Data Centres are growing in size and processing capacity to manage the increased size and flow of data. As the Clouds they are hosting are evolving to perform more processing intensive tasks and in particular the ones involved in artificial intelligence -training, inferencing- there is an interest for chips that can better support these activities.
Here lays the interest for the Intel announcement of the 3rd generation of Xeon chips specifically designed to support AI intensive processing.
These chips are using 14nm technology, hence they are not at the leading edge with respect to etching dimension but support bfloat 16 instruction set (this is a compact numeric format that can achieve processing effectiveness similar at FP32 at a lower price point, something that is very important in large data centres). The chips are offered with 16 or 28 cores supporting multi-thread (32 to 56 respectively) and embed 38.5MB of cache memory (don’t be fooled into thinking that 38MB is peanuts, that is a cache memory used to support processing inside the chip, not a data storage) and their clock speed of 2.3 GHz can be boosted as needed up to 4.3 GHz. This is a very nice feature since one of the big cost of data centre is powering the chips (electricity) and cooling them (air conditioning). By keeping the clock speed at 2.3 GHz and boosting it only when there is a need the power consumption and dissipation is greatly reduced making them ideal for use in data centres.
Also inportant is the software backward compatibility (not 100% but according to Intel only minor changes will be required).
This news is interesting because it shows the trend towards commoditising artificial intelligence that becomes an integral feature of data centres.