Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
T
tnl-dev
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Deploy
Releases
Model registry
Analyze
Value stream analytics
Contributor analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
This is an archived project. Repository and other project resources are read-only.
Show more breadcrumbs
TNL
tnl-dev
Graph
7d389df70a3fce417782c95dd9bfe2ef28aee131
Select Git revision
No matching results
An error occurred while fetching branches. Retry the search.
An error occurred while fetching tags. Retry the search.
You can move around the graph by using the arrow keys.
Begin with the selected commit
Created with Raphaël 2.2.0
23
Aug
22
17
16
14
10
9
8
7
6
5
4
3
2
1
31
Jul
30
29
28
27
26
25
16
15
14
13
12
11
10
9
7
5
4
2
1
25
Jun
24
20
19
18
17
14
24
May
5
29
Apr
26
22
21
20
19
16
15
14
13
12
11
6
5
3
1
31
Mar
30
29
28
27
25
8
2
1
15
Feb
14
10
finished merge, L-F and S-W operators changed to accept classes as equation's RHS
resolved work for host /cuda in differential operators
removed solitare flow solvers for Euler and Navier-Stokes equations
added solver containing Steger-Warming, Van Leer and Lax Friedrich in one for Euler and Navier-Stokes
Merge branch 'JK/ci' into 'develop'
Documentation: fixed branch name for CI builds
Merge branch 'JK/ci' into 'develop'
Documentation: added git commit id to the header
CI: split build and deploy documentation, use artifacts
CI: use the dummy build job only for merge requests
Merge branch 'doc-deploy' into 'develop'
CI: automatically skip builds for commits without any changes in the source files
Setup for automatic documentation deployment
Merge branch 'JK/multireduction' into 'develop'
Implemented distributed prefix-sum
PrefixSum: separate first and second phase for OpenMP implementation and expose performFirstPhase and performSecondPhase methods
CUDA prefix-sum: separated the implementation of the first and second phase
CUDA prefix-sum: moved gridShift from the first phase to the second phase
Replaced static member variables in CudaPrefixSumKernelLauncher with static getters
Added default stream synchronizations after kernel launches in CudaPrefixSumKernel.h
Removed volatile reduction from PrefixSum and updated the normal reduction operation
Added prefix-sum to BLAS benchmarks
Implemented parallel prefix-sum with OpenMP
Benchmarks: compute sample standard deviation of the measured computation times
Removed timing parameter from benchmarks
Benchmarks: added scalar multiplication with BLAS
Optimized OpenMP thread counts for reduction and multireduction
Ugly workaround for nvcc's stupid modification of `new` expressions
Replaced custom lambda functions with instances of STL types where possible
Changed reduction operation to use functions with `return a + b` instead of `a += b`
Removed VectorOperations class which is now useless
Removed ReductionOperations.h
Removed volatile reduction completely
Found a way to avoid using volatile in CUDA reduction: __syncwarp()
Rewritten multireduction with lambda functions
Style changes in the code for reduction
Removed reduction and multireduction declarations for MIC
Execute tests in parallel
Tests: forced using a unique file name in each test
Disabled long double in tnl-diff, tnl-init, tnl-lattice-init and tnl-view
Loading