Intel disclosed new technical details about its next -generation Xeon Phi processor to a small group of technology analysts-journalists at one of its Hillsboro sites this week. First revealed last year,
the “Knights Landing” chip is expected to ship later this year, however
Intel showed pre-production parts parts running in a Hillsboro test
lab. The briefing included a whiteboard design walkthrough by Avinash
Sodani, Knights Landing Chief Architect at Intel, a product update by
Hugo Saleh, Marketing Director of Intel’s Technical Computing Group, an
interactive technical Q&A and a lab demo of a Knights Landing system
running on an Intel reference-design system. The architectural
differences between Knights Landing and prior Phi products is
substantial and should translate to much easier adoption and a broader
set of commercial applications, including areas like deep learning,
genomics, databases and general purpose containerized (Docker)
workloads. Unlike more specialized processors, Intel describes Knights
Landing as taking a “holistic approach” to new breakthrough
applications.
Unlike the current generation Phi design, which operates as a
coprocessor, Knights Landing incorporates x86 cores and can directly
boot and run standard operating systems and application code without
recompilation. The test system, which resembled a standard server board,
with socketed CPU and memory modules was running a stock Linux
distribution. This is made possible by the inclusion of a modified
version of the Atom Silvermont x86 cores
as part of each Knights Landing ’tile’: the chip’s basic design unit
consisting of dual x86 and vector execution units alongside cache memory
and intra-tile mesh communication circuitry. Each multi-chip package
includes the a processor with 30 or more (rumored up to 36) tiles and
eight high-speed memory chips. Although it wouldn’t offer specifics,
Intel said the on-package memory, totalling 16GB, is made by Micron with
custom I/O circuitry and might be a variant of Micron’s announced, but not yet shipping Hybrid Memory Cube, but this wasn’t confirmed. The high-speed memory is conceptually similar to the DDR5 devices used on GPUs like NVIDIA's NVDA -0.58% Tesla.
Intel is long on device-level technology details, but short on
specifics regarding the types of business problems, applications and
users Phi targets. So far, Phi has been used by HPC applications like
scientific simulations, modeling and design, with product announcements
often happening during supercomputing conferences. Yet as NVIDIA
demonstrated at GTC,, the opportunities for compute hardware supporting
wide, vector operations built for highly parallelized algorithms,
whether general purpose GPUs or now Xeon Phi, extends far beyond the
supercomputer niche. As I previously discussed,
NVIDIA has morphed its GPU architecture into a powerful engine for
calculating deep learning neural network algorithms that is now being
applied to accelerate SQL database operations and analytics.
The internals of a GPU and Xeon Phi are much different, however both
share many common traits: dozens (Phi) to thousands (GPU) of lower
performance (Phi) to relatively simple cores (GPU), vector processing
units and very high-speed local memory and buses. The creative
exploitation of this capability isn’t limited to scientific or graphics
calculations and seems limited only by the imagination of developers.
But on this front, Intel didn’t have encouraging news. Although Saleh
mentioned deep learning as a potential application, he couldn’t point to
specific customers or researchers using Phi for anything but
traditional scientific HPC problems.
Unlike NVIDIA, which announced the price and availability of a Titan X development box designed
for researchers exploring GPU applications to deep learning, Intel
wouldn’t share details about when OEM partners will have Knights Landing
systems, whether any ould be sized and priced for individual
developers, nor a timeline for what Saleh characterized as a Phi
developer program, which will include access to hardware, updated
software tools and training materials. But developer resources, outreach
and evangelism are critical to any success Phi is to have, particularly
since the hardware itself represents a substantial departure from the
standard x86 microarchitecture, and here NVIDIA has a big head start.
Yet thinking of Phi as an alternative GPU or vector processor is
wrong since it’s actually a hybrid that includes dozens of full-fledged
64-bit x86 cores. This unique design, if used right, could significantly
accelerate certain parallelizable application categories besides HPC
simulation like deep learning and data analytics that use vector
calculations, while offering drop-in computational offload for standard
x86 code.
Read more:
http://www.forbes.com/sites/kurtmarko/2015/03/25/intels-deep-learning-play/
No comments:
Post a Comment