Compiler benchmark: gcc clang pcc tcc

Compiler benchmark: gcc clang pcc tcc

CompilerDeathMatch-x4-64bit

This is a collection of compiler benchmarks.

Test results:

  1. Timed Apache Compilation 2.2.11
  2. Timed PHP Compilation 5.2.9
  3. Timed ImageMagick Compilation 6.6.3-4
  4. Apache Benchmark 2.2.11 — Static Web Page Serving
  5. 7-Zip Compression 9.04 — Compress Speed Test
  6. LAME MP3 Encoding 3.98.2 — WAV To MP3
  7. Gcrypt Library 1.4.4 — CAMELLIA256-ECB Cipher
  8. OpenSSL 1.0.0a — RSA 4096-bit Performance
  9. C-Ray 1.1 — Total Time
  10. Bullet Physics Engine 2.75 — Test: 3000 Fall
  11. Bullet Physics Engine 2.75 — Test: 1000 Stack
  12. Bullet Physics Engine 2.75 — Test: 1000 Convex
  13. Bullet Physics Engine 2.75 — Test: Convex Trimesh
  14. John The Ripper 1.7.3.1 — Test: Blowfish
  15. Timed HMMer Search 2.3.2 — Pfam Database Search

NVIDIA GeForce GTX 460 – Der Jäger stellt sich vor – Tests …

NVIDIA GeForce GTX 460 – Der Jäger stellt sich vor – Tests …
Beide Modelle agieren mit hoher Leistung in der 200-US-Dollar-Region. Ganz offensichtlich haben die Boardpartner bei diesen Grafikkarten schon unmittelbar zum Start hin großen Spielraum. Referenzmodelle bemustert NVIDIA nämlich nicht und die Boardpartner rühmen sich, von Anfang an eigene Designs und übertaktete Versionen zu zeigen.Wir dürfen heute gleich zwei Top-Modelle der Hersteller Gainward und MSI präsentieren. Die Gainward GeForce GTX 460 Golden Sample mit 1024 MByte und der MSI N460GTX Cyclone mit 768 MByte Hauptspeicher. Beide Grafikkarten sind von Hause aus schon übertaktet.

Andrew Corrigan: Porting Large Fortran Codebases to CUDA

Andrew Corrigan: Porting Large Fortran Codebases to CUDA

A converter was developed to automatically port the code FEFLO to GPUs.

Full GPU performance
• Port ~1 million lines of code (~11,000 parallel loops)
• Continue development in Fortran using established coding practices.
• A single, unified codebase.

Using a Python script
• O(1000) line Python script based on FParser
• Developed in a few months.
• Generates an optimized, running code.
• Does much more than translate loops in isolation.
• Generates CUDA kernels from existing OpenMP and vector loops.
• Tracks array usage across the entire code.
• By far the most difficult task.
• Many other tasks.

CULA

CULA

CULA is a GPU-accelerated linear algebra library that utilizes the NVIDIA CUDA parallel computing architecture to dramatically improve the computation speed of sophisticated mathematics.

CULAtools™ is EM Photonics’ product family comprised of CULA™ Basic, Premium, and Commercial. CULA is our GPU-accelerated implementation of the LAPACK numerical linear algebra library, containing several of the most popular LAPACK functions. After developing accelerated linear algebra solvers since 2004 for our clients, EM Photonics partnered with NASA Ames Research Center in 2007 to extend and unify these libraries into a single, GPU-accelerated package. Through a partnership with NVIDIA®, we focused on developing a commercially available implementation of accelerated LAPACK routines. Our primary goal is to help a wide range of users experience computational performance previously available only on supercomputers. By leveraging NVIDIA’s CUDA™ architecture, CULA provides users linear algebra functions with unsurpassed performance.

swan – Multiscale Laboratory

swan – Multiscale Laboratory
What is it?Swan is a small tool that aids the reversible conversion of existing CUDA codebases to OpenCL. It does several useful things:

Translates CUDA kernel source-code to OpenCL.

Provides a common API that abstracts both CUDA and OpenCL runtimes.

Preserves the convenience of the CUDA <<>> kernel launch syntax by generating C source-code for kernel entry-point functions.

It can also be usefully used for compiling and managing kernels written directly for OpenCL.

Why might you want it?

Possible uses include:

Evaluating OpenCL performance of an existing CUDA code.

Maintaining a dual-target OpenCL and CUDA code.

Reducing dependence on NVCC when compiling host code.Support multiple CUDA compute capabilities in a single binary

A runtime library for managing OpenCL kernels for new development

NAMD – Scalable Molecular Dynamics

NAMD – Scalable Molecular Dynamics
NAMD, recipient of a 2002 Gordon Bell Award, is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. Based on Charm++ parallel objects, NAMD scales to hundreds of processors on high-end parallel platforms and tens of processors on commodity clusters using gigabit ethernet. NAMD uses the popular molecular graphics program VMD for simulation setup and trajectory analysis, but is also file-compatible with AMBER, CHARMM, and X-PLOR. NAMD is distributed free of charge with source code. You can build NAMD yourself or download binaries for a wide variety of platforms. Our tutorials show you how to use NAMD and VMD for biomolecular modeling.The 2005 reference paper Scalable molecular dynamics with NAMD has over 1000 citations as of March 2010.