Intel_C++_Compiler

Intel C++ Compiler

Intel C++ Compiler

Compiler


Intel oneAPI DPC++/C++ Compiler and Intel C++ Compiler Classic (deprecated icc and icl[3] is in Intel OneAPI HPC toolkit) are Intel’s C, C++, SYCL, and Data Parallel C++ (DPC++) compilers for Intel processor-based systems, available for Windows, Linux, and macOS operating systems.[4]

Quick Facts Other names, Developer(s) ...
Quick Facts Other names, Developer(s) ...

Overview

Intel oneAPI DPC++/C++ Compiler is available for Windows and Linux and supports compiling C, C++, SYCL, and Data Parallel C++ (DPC++) source, targeting Intel IA-32, Intel 64 (aka x86-64), Core, Xeon, and Xeon Scalable processors, as well as GPUs including Intel Processor Graphics Gen9 and above, Intel Xe architecture, and Intel Programmable Acceleration Card with Intel Arria 10 GX FPGA.[5] Like Intel C++ Compiler Classic, it also supports the Microsoft Visual Studio and Eclipse IDE development environments, and supports threading via Intel oneAPI Threading Building Blocks, OpenMP, and native threads.

DPC++[6][7] builds on the SYCL specification from The Khronos Group. It is designed to allow developers to reuse code across hardware targets (CPUs and accelerators such as GPUs and FPGAs) and perform custom tuning for a specific accelerator. DPC++ comprises C++17 and SYCL language features and incorporates open-source community extensions that make SYCL easier to use. Many of these extensions were adopted by the SYCL 2020 provisional specification[8] including unified shared memory, group algorithms, and sub-groups.

Intel announced in August 2021 the complete adoption of LLVM for faster build times and benefits from supporting the latest C++ standards.[9]

Intel C++ Compiler Classic is available for Windows, Linux, and macOS and supports compiling C and C++ source, targeting Intel IA-32, Intel 64 (x86-64), Core, Xeon, and Xeon Scalable processors.[5] It supports the Microsoft Visual Studio and Eclipse IDE development environments. Intel C++ Compiler Classic supports threading via Intel oneAPI Threading Building Blocks, OpenMP, and native threads.

Architectures

According to Intel,[10] starting with the 2023.0 release, Intel oneAPI DPC++/C++ Compiler supports all current Intel general-purpose x86-64 CPUs and GPUs including:

  • Processors:
    • Legacy Intel IA-32 and Intel 64 (x86-64) processors
    • Intel Core processors
    • Intel Xeon processor family
    • Intel Xeon Scalable processors
    • Intel Xeon Processor Max Series
  • GPUs:
    • Intel Processor Graphics Gen9 and above
    • Intel Xe architecture
    • Intel Programmable Acceleration Card with Intel 10 GX FPGA
    • Intel Data Center GPUs including Flex Series and Max Series
  • Intel FPGAs

Intel C++ Compiler Classic targets general-purpose Intel x86-64 architecture CPUs including:[5]

  • Legacy Intel IA-32 and Intel 64 (x86-64) processors
  • Intel Core processors
  • Intel Xeon processor family
  • Intel Xeon Scalable processors

Toolkits

The Intel oneAPI DPC++/C++ Compiler is available either as a standalone component[11] or as part of the Intel oneAPI Base Toolkit, Intel oneAPI HPC Toolkit, and Intel oneAPI IoT Toolkit.[5]

The Intel C++ Compiler Classic is available either as a standalone component[12] or as part of the Intel oneAPI Base Toolkit.[5]

Debugging

The Intel compiler provides debugging information that is standard for the common debuggers (DWARF 2 on Linux, similar to gdb, and COFF for Windows). The flags to compile with debugging information are /Zi on Windows and -g on Linux. Debugging is done on Windows using the Visual Studio debugger and, on Linux, using gdb.

While the Intel compiler can generate a gprof compatible profiling output, Intel also provides a kernel level, system-wide statistical profiler called Intel VTune Profiler. VTune can be used from a command line or through an included GUI on Linux or Windows. It can also be integrated into Visual Studio on Windows, or Eclipse on Linux). In addition to the VTune profiler, there is Intel Advisor that specializes in vectorization optimization, offload modeling, flow graph design and tools for threading design and prototyping.

Intel also offers a tool for memory and threading error detection called Intel Inspector XE. Regarding memory errors, it helps detect memory leaks, memory corruption, allocation/de-allocation of API mismatches and inconsistent memory API usage. Regarding threading errors, it helps detect data races (both heap and stack), deadlocks and thread and synch API errors.

Support for non-Intel processors

Previous versions of Intel’s C/C++ compilers have been criticized for optimizing less aggressively for non-Intel processors; for example, Steve Westfield wrote in a 2005 article at the AMD website:[13]

Intel 8.1 C/C++ compiler uses the flag -xN (for Linux) or -QxN (for Windows) to take advantage of the SSE2 extensions. For SSE3, the compiler switch is -xP (for Linux) and -QxP (for Windows). ... With the -xN/-QxN and -xP/-QxP flags set, it checks the processor vendor string—and if it's not "GenuineIntel", it stops execution without even checking the feature flags. Ouch!

The Danish developer and scholar Agner Fog wrote in 2009:[14]

The Intel compiler and several different Intel function libraries have suboptimal performance on AMD and VIA processors. The reason is that the compiler or library can make multiple versions of a piece of code, each optimized for a certain processor and instruction set, for example SSE2, SSE3, etc. The system includes a function that detects which type of CPU it is running on and chooses the optimal code path for that CPU. This is called a CPU dispatcher. However, the Intel CPU dispatcher does not only check which instruction set is supported by the CPU, it also checks the vendor ID string. If the vendor string is "GenuineIntel" then it uses the optimal code path. If the CPU is not from Intel then, in most cases, it will run the slowest possible version of the code, even if the CPU is fully compatible with a better version.

This vendor-specific CPU dispatching may potentially impact the performance of software built with an Intel compiler or an Intel function library on non-Intel processors, possibly without the programmer’s knowledge. This has allegedly led to misleading benchmarks,[14] including one incident when changing the CPUID of a VIA Nano significantly improved results.[15] In November 2009, AMD and Intel reached a legal settlement over this and related issues,[16] and in late 2010, AMD settled a US Federal Trade Commission antitrust investigation against Intel.[17]

The FTC settlement included a disclosure provision where Intel must:[18]

publish clearly that its compiler discriminates against non-Intel processors (such as AMD's designs), not fully utilizing their features and producing inferior code.

In compliance with this ruling, Intel added disclaimers to its compiler documentation:[19]

Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice.

As late as 2013, an article in The Register alleged that the object code produced by the Intel compiler for the AnTuTu Mobile Benchmark omitted portions of the benchmark which showed increased performance compared to ARM platforms.[20]

Release history

The following lists versions of the Intel C++ Compiler since 1999:[21]

More information Compiler version, Release date ...

See also


References

  1. Intel Corporation (2022-11-02). "Intel® oneAPI DPC++/C++ Compiler". software.intel.com. Intel. Retrieved 2022-12-01.
  2. Intel Corporation (2022-11-02). "Intel® C++ Compiler Classic". software.intel.com. Intel. Retrieved 2022-12-01.
  3. "DEPRECATION NOTICE: Intel® C++ Compiler Classic". community.intel.com. 2022-09-01. Retrieved 2023-02-13.
  4. Intel (2021). "Intel oneAPI DPC++/C++ Compiler". Intel.com. Intel. Retrieved 2021-02-09.
  5. Intel Corporation (2021). "Intel® oneAPI DPC++/C++ Compiler". software.intel.com. Intel. Retrieved 2021-02-09.
  6. Team, Editorial (2019-12-16). "Heterogeneous Computing Programming: oneAPI and Data Parallel C++". insideBIGDATA. Retrieved 2020-12-17.
  7. Intel Corporation (November 30, 2022). "Intel oneAPI 2023 Release: Preview the Tools". www.intel.com. Intel. Retrieved 2022-12-01.
  8. Intel Corporation (2020-12-16). "Intel® oneAPI DPC++/C++ Compiler". software.intel.com. Intel. Retrieved 2021-02-09.
  9. Intel Corporation (2020-12-16). "Intel® C++ Compiler Classic". software.intel.com. Intel. Retrieved 2021-02-09.
  10. Hruska, Joel (29 July 2008). "Low-end grudge match: Nano vs. Atom". Ars Technica.
  11. "Settlement agreement" (PDF). download.intel.com.
  12. "Intel and U.S. Federal Trade Commission Reach Tentative Settlement". Newsroom.intel.com. 2010-08-04. Retrieved 2012-10-13.
  13. "Optimization Notice". Intel Corporation. Retrieved 11 December 2013.
  14. "Intel® C++ Compiler Release Notes and New Features". Intel Corporation. Retrieved 27 April 2021.
  15. This note is attached to the release in which Cilk Plus was introduced. This URL points to current documentation: http://software.intel.com/en-us/intel-composer-xe/

Share this article:

This article uses material from the Wikipedia article Intel_C++_Compiler, and is written by contributors. Text is available under a CC BY-SA 4.0 International License; additional terms may apply. Images, videos and audio are available under their respective licenses.