Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
hpc-compendium
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Deploy
Releases
Package registry
Container Registry
Model registry
Operate
Terraform modules
Monitor
Incidents
Service Desk
Analyze
Value stream analytics
Contributor analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
ZIH
hpcsupport
hpc-compendium
Commits
cd16c1fa
Commit
cd16c1fa
authored
3 years ago
by
Martin Schroschk
Browse files
Options
Downloads
Patches
Plain Diff
Resolve review
parent
45f2dc47
No related branches found
No related tags found
3 merge requests
!322
Merge preview into main
,
!319
Merge preview into main
,
!249
review libraries.md
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
doc.zih.tu-dresden.de/docs/software/math_libraries.md
+8
-8
8 additions, 8 deletions
doc.zih.tu-dresden.de/docs/software/math_libraries.md
with
8 additions
and
8 deletions
doc.zih.tu-dresden.de/docs/software/math_libraries.md
+
8
−
8
View file @
cd16c1fa
...
...
@@ -26,15 +26,15 @@ factorization, and eigenvalue problems.
The
[
Scalable Linear Algebra PACKage
](
https://www.netlib.org/scalapack
)
(
ScaLAPACK
)
takes the
idea of high-performance linear algebra routines to parallel distributed memory machines. It offers
functionality to solve
s
dense and banded linear systems, least squares problems, eigenvalue
functionality to solve dense and banded linear systems, least squares problems, eigenvalue
problems, and singular value problems.
<!--There is also an [optimized implementation](https://developer.amd.com/amd-aocl/scalapack/) addressing-->
<!--AMD architectures.-->
Many concrete implementations, often tuned and optimized for specific hardware architectures, have
been developed over the last decades. The two hardware vendors Intel and AMD each offer
a
own math
library -
[
Intel MKL
](
#math-kernel-library-mkl
)
and
[
AOCL
](
#amd-optimizing-cpu-libraries-aocl
)
)
.
been developed over the last decades. The two hardware vendors Intel and AMD each offer
their
own math
library -
[
Intel MKL
](
#math-kernel-library-mkl
)
and
[
AOCL
](
#amd-optimizing-cpu-libraries-aocl
)
.
Both libraries are worth to consider from a users point of view, since they provide extensive math
functionality ranging from BLAS and LAPACK to random number generators and Fast Fourier
Transformation with consistent interfaces and the "promises" to be highly tuned and optimized and
...
...
@@ -48,7 +48,7 @@ continuously developed further.
## AMD Optimizing CPU Libraries (AOCL)
AMD Optimizing CPU Libraries
(AOCL)
(https://developer.amd.com/amd-aocl/) is a set of numerical
[
AMD Optimizing CPU Libraries
]
(
https://developer.amd.com/amd-aocl/
)
(
AOCL
)
is a set of numerical
libraries tuned specifically for AMD EPYC processor family. AOCL offers linear algebra libraries
(
[
BLIS
](
https://developer.amd.com/amd-cpu-libraries/blas-library/
)
,
[
libFLAME
](
https://developer.amd.com/amd-cpu-libraries/blas-library/#libflame
)
,
...
...
@@ -59,7 +59,7 @@ libraries tuned specifically for AMD EPYC processor family. AOCL offers linear a
as well as
[
AMD Random Number Generator Library
](
https://developer.amd.com/amd-cpu-libraries/rng-library/
)
and
[
AMD Secure RNG Library(https://developer.amd.com/amd-cpu-libraries/rng-library/#securerng).
[
AMD Secure RNG Libra
a
ry
]
(
https://developer.amd.com/amd-cpu-libraries/rng-library/#securerng
)
.
## Math Kernel Library (MKL)
...
...
@@ -92,7 +92,7 @@ For linker flag combinations, we highly recommand the
GPU implementations of math functions and routines are often much faster compared to CPU
implementations. This also holds for basic routines from BLAS and LAPACK. You should consider using
this
GPU implementations
first
in order to obtain better performance.
GPU implementations in order to obtain better performance.
There are several math libraries for Nvidia GPUs, e.g.
...
...
@@ -101,8 +101,8 @@ There are several math libraries for Nvidia GPUs, e.g.
-
[
cuSPARSE
](
https://developer.nvidia.com/cusparse
)
(
sparse
matrix library)
-
[
cuFFT
](
https://developer.nvidia.com/cufft
)
[
This webpage
](
https://developer.nvidia.com/gpu-accelerated-libraries#linear-algebra
)
provides a
comprehensive overview and starting point.
Nvidia
provides a
[
comprehensive overview and starting point
](
https://developer.nvidia.com/gpu-accelerated-libraries#linear-algebra
)
.
### MAGMA
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment