PmWiki
Results

Hybrid MIMD/SIMD high order DGTD solver

Results.HPC History

Show minor edits - Show changes to output

May 06, 2015, at 02:29 PM by 138.96.201.175 -
Changed lines 26-28 from:
(:cellnr align='center':) %width=550px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=600px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
(:cellnr align='center':) %width=600px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3-new.jpg
(:cellnr align='center':) %width=600px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4-new.jpg
May 06, 2015, at 02:29 PM by 138.96.201.175 -
Changed line 26 from:
(:cellnr align='center':) %width=500px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
to:
(:cellnr align='center':) %width=550px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
May 06, 2015, at 02:27 PM by 138.96.201.175 -
Changed line 26 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
to:
(:cellnr align='center':) %width=500px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
May 06, 2015, at 02:26 PM by 138.96.201.175 -
Changed line 26 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.png
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2-new.jpg
May 06, 2015, at 02:23 PM by 138.96.201.175 -
Changed line 26 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.png
April 30, 2015, at 11:07 AM by 134.94.123.1 -
Changed lines 16-18 from:
We conducted a parallel performance evaluation in terms of strong scalability analysis on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh is made of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.

to:
We selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh is made of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
Changed lines 19-21 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
Deleted lines 20-25:
%center% Strong scalability analysis of the DGTD solver


(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
(:tableend:)
Added lines 22-30:

We conducted a parallel performance evaluation in terms of strong scalability analysis on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.

(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
(:tableend:)
%center% Strong scalability analysis of the DGTD solver
April 30, 2015, at 11:04 AM by 134.94.123.1 -
Changed lines 20-21 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg 
(:cellnr align='center':)
%width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
April 30, 2015, at 11:03 AM by 134.94.123.1 -
Changed lines 20-23 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell   align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:tableend:)
(:table border='0' width='100%' align='center' cellspacing='1px':)
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 30, 2015, at 11:03 AM by 134.94.123.1 -
Changed lines 21-23 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
to:
(:cell  align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:tableend:)
(:table border='0' width='100%' align='center' cellspacing='1px':)
April 30, 2015, at 11:02 AM by 134.94.123.1 -
Changed line 28 from:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
to:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
April 30, 2015, at 11:01 AM by 134.94.123.1 -
Changed line 28 from:
(:cellnr align='center':) %width=300px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
to:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
Changed line 30 from:
%center% Contour lines of the amplitude of the DFT of '''E''''_z_'
to:
%center% Contour lines of the amplitude of the DFT of '''E'''
April 30, 2015, at 11:01 AM by 134.94.123.1 -
Changed lines 28-29 from:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-1.jpg
(:cell  align='center':) %width=300px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-2.jpg
to:
(:cellnr align='center':) %width=300px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
April 29, 2015, at 04:15 PM by 134.94.123.1 -
Deleted line 4:
Changed lines 18-19 from:
(:linebreaks:)
to:
Deleted line 25:
(:linebreaks:)
April 29, 2015, at 04:15 PM by 134.94.123.1 -
Changed line 5 from:
>>frame bgcolor=white<<
to:
Changed lines 7-9 from:
>><<

>>frame bgcolor=white<<
to:
Changed lines 16-18 from:
>><<

>>frame bgcolor=white<<
to:
Deleted line 17:
>><<
April 29, 2015, at 04:12 PM by 134.94.123.1 -
Changed line 9 from:
>>frame<<
to:
>>frame bgcolor=white<<
Changed line 20 from:
>>frame<<
to:
>>frame bgcolor=white<<
April 29, 2015, at 04:12 PM by 134.94.123.1 -
Changed line 5 from:
>>frame<<
to:
>>frame bgcolor=white<<
April 29, 2015, at 04:07 PM by 134.94.123.1 -
Changed line 21 from:
We conducted a parallel performance evaluation in terms of strong scalability analysis. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh is made of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
to:
We conducted a parallel performance evaluation in terms of strong scalability analysis on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh is made of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
Deleted lines 22-25:

>>frame<<
The strong scalability analysis has been conducted on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.
>><<
April 29, 2015, at 04:05 PM by 134.94.123.1 -
Changed lines 25-26 from:
The strong scalability analysis has been conducted on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.
to:
The strong scalability analysis has been conducted on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.
>><<

Deleted line 43:
>><<
April 29, 2015, at 04:05 PM by 134.94.123.1 -
Changed lines 20-21 from:
(:linebreaks:)
to:
>>frame<<
Changed lines 22-26 from:
to:
>><<

>>frame<<
The strong scalability analysis has been conducted on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.

Deleted lines 28-31:
The strong scalability analysis has been conducted on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.

(:linebreaks:)

Added line 43:
>><<
April 29, 2015, at 04:03 PM by 134.94.123.1 -
Added line 7:
>><<
April 29, 2015, at 04:03 PM by 134.94.123.1 -
Changed lines 6-8 from:
This study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452).>><<

(:linebreaks:)
to:
This study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452).
April 29, 2015, at 04:03 PM by 134.94.123.1 -
Changed lines 6-8 from:
This study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452).
>><<
to:
This study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452).>><<
Changed line 10 from:
>>frame bgcolor='white'<<
to:
>>frame<<
April 29, 2015, at 04:02 PM by 134.94.123.1 -
April 29, 2015, at 04:02 PM by 134.94.123.1 -
Changed lines 5-6 from:
>>frame bgcolor='white'<<
This
study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452). 
to:
>>frame<<
This
study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452).
April 29, 2015, at 04:02 PM by 134.94.123.1 -
Added line 5:
>>frame bgcolor='white'<<
Added line 7:
>><<
April 29, 2015, at 03:26 PM by 134.94.123.1 -
April 29, 2015, at 03:26 PM by 134.94.123.1 -
Deleted line 9:
Deleted line 17:
April 29, 2015, at 03:25 PM by 134.94.123.1 -
Changed lines 9-10 from:
>frame bgcolor='white'<<
to:
>>frame bgcolor='white'<<
Added line 19:
April 29, 2015, at 03:25 PM by 134.94.123.1 -
Added line 9:
>frame bgcolor='white'<<
Added line 18:
>><<
April 29, 2015, at 03:14 PM by 134.94.123.1 -
Changed line 24 from:
The strong scalability analysis has been conducted on the thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.
to:
The strong scalability analysis has been conducted on thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.
April 29, 2015, at 03:13 PM by 134.94.123.1 -
Changed line 20 from:
We conducted a parallel performance evaluation in terms of strong scalability analysis. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh consists of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
to:
We conducted a parallel performance evaluation in terms of strong scalability analysis. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh is made of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
April 29, 2015, at 03:02 PM by 134.94.123.1 -
Changed line 20 from:
We conducted a parallel performance evaluation in terms of strong scalability analysis. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists in nanosphere embedded in vacuum. The constructed tetrahedral mesh consists of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
to:
We conducted a parallel performance evaluation in terms of strong scalability analysis. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists of a nanosphere chain embedded in vacuum. The constructed tetrahedral mesh consists of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
April 29, 2015, at 03:01 PM by 134.94.123.1 -
Changed line 16 from:
Indeed, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing for  harnessing the processing capabilities of manycore CPUs or accelerator chips. We are concerned here with the study of a hybrid coarse grain/fine grain parallelization of a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation  of light interaction with nanometer scale metallic structures.
to:
Indeed, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing for  harnessing the processing capabilities of manycore CPUs or accelerator chips. We are concerned here with the study of a hybrid coarse grain/fine grain parallelization strategy for a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation  of light interaction with nanometer scale metallic structures.
April 29, 2015, at 03:00 PM by 134.94.123.1 -
Changed line 16 from:
Indeed, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing fro exploiting the processing capabilities of manycore CPUs or accelerator chips. We are concerned here with the study of a hybrid coarse grain/fine grain parallelization of a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation  of light interaction with nanometer scale metallic structures.
to:
Indeed, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing for  harnessing the processing capabilities of manycore CPUs or accelerator chips. We are concerned here with the study of a hybrid coarse grain/fine grain parallelization of a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation  of light interaction with nanometer scale metallic structures.
April 29, 2015, at 02:59 PM by 134.94.123.1 -
Changed lines 16-17 from:
Indeed, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing fro exploiting the processing capabilities of manycore CPUs or accelerator chips.
We are concerned here with the study of a hybrid coarse grain/fine grain parallelization of a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation  of light interaction with nanometer scale metallic structures.
to:
Indeed, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing fro exploiting the processing capabilities of manycore CPUs or accelerator chips. We are concerned here with the study of a hybrid coarse grain/fine grain parallelization of a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation  of light interaction with nanometer scale metallic structures.
April 29, 2015, at 02:59 PM by 134.94.123.1 -
Changed lines 14-16 from:
* It is naturally adapted to parallel computing.  As long as an explicit time integration scheme is used, the DGTD method is easily parallelized.  Moreover, the compact nature of method is in favor of high computation  to communication ratio  especially when the interpolation order is increased.

In this study, we focus on the last of these features and develop a hybrid coarse grain/fine grain parallelization of a high order DGTD solver formulated on unstructured tetrahedral meshes for the simulation
of light interaction with nanometer scale metallic structures.
to:
* It is naturally adapted to parallel computing.  As long as an explicit time integration scheme is used, the DGTD method is easily parallelized. 

Indeed
, the compact nature of method (the polynomial interpolation of the physical field is performed at the element level) is particularly appealing fro exploiting the processing capabilities of manycore CPUs or accelerator chips.
We are concerned here with the study of a hybrid coarse grain/fine grain parallelization of a high order DGTD solver for the system of Maxwell equations coupled to a physical dispersion model. Practical modeling settings of interest to our study are the system of Maxwell equations coupled to a Debye dispersion model for the simulation of microwave interaction with biologocal tissues, and the system of Maxwell equations coupled to a Drude dispersion model for the simulation
of light interaction with nanometer scale metallic structures.
April 29, 2015, at 01:41 PM by 134.94.123.1 -
Changed lines 29-31 from:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 29, 2015, at 01:40 PM by 134.94.123.1 -
Changed lines 29-33 from:
(:cellnr align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:tableend:)
(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=365px
% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 27, 2015, at 10:57 AM by 138.96.201.175 -
Changed line 41 from:
(:cell  align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-2.jpg
to:
(:cell  align='center':) %width=300px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-2.jpg
April 27, 2015, at 10:57 AM by 138.96.201.175 -
Changed line 41 from:
(:cell  align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-2.jpg
to:
(:cell  align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-2.jpg
April 27, 2015, at 10:56 AM by 138.96.201.175 -
Added line 41:
(:cell  align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-2.jpg
April 27, 2015, at 10:54 AM by 138.96.201.175 -
Changed line 40 from:
(:cellnr align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-1.jpg
to:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-1.jpg
April 27, 2015, at 10:54 AM by 138.96.201.175 -
Added lines 36-42:

(:linebreaks:)

(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/Ez-1.jpg
(:tableend:)
%center% Contour lines of the amplitude of the DFT of '''E''''_z_'
April 25, 2015, at 01:16 PM by 82.228.254.112 -
Added lines 38-46:

!!!Related publications

(:linebreaks:)

S. Lanteri, R. Léger, C. Scheid,J. Viquerat, T. Cabel and G. Hautreux\\
Hybrid MIMD/SIMD high order DGTD solver for the numerical modeling of light/matter interaction on the nanoscale\\
%newwin% [[http://www.prace-ri.eu/IMG/pdf/WP207.pdf |  PRACE White Paper (2015)]]\\

April 25, 2015, at 01:09 PM by 82.228.254.112 -
Changed lines 29-30 from:
(:cellnr align='center':) %width=375px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=375px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
to:
(:cellnr align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
Changed line 33 from:
(:cellnr align='center':) %width=375px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=365px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 25, 2015, at 01:06 PM by 82.228.254.112 -
Changed lines 29-30 from:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
to:
(:cellnr align='center':) %width=375px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=375px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
Changed line 33 from:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=375px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 25, 2015, at 01:06 PM by 82.228.254.112 -
Changed lines 29-30 from:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
to:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cell  align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
Changed line 33 from:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
Added line 35:
%center% Strong scalability analysis of the DGTD solver
April 25, 2015, at 01:03 PM by 82.228.254.112 -
Added lines 24-27:
The strong scalability analysis has been conducted on the thin nodes of the %newwin% [[http://www.prace-ri.eu/prace-resource | Curie system]]. Each run has been made considering 8 OpenMP threads per socket and 2 sockets per node.

(:linebreaks:)

Added lines 31-32:
(:tableend:)
(:table border='0' width='100%' align='center' cellspacing='1px':)
April 25, 2015, at 12:56 PM by 82.228.254.112 -
Changed line 26 from:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
to:
(:cell  align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
April 25, 2015, at 12:54 PM by 82.228.254.112 -
Changed lines 25-27 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
to:
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=400px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 25, 2015, at 12:48 PM by 82.228.254.112 -
Changed lines 20-24 from:
(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=350px% http://www-sop
.inria.fr/nachos/pics/results/hpc/Yguide-1.png
(:cell  align='center':) %width=350px% http://www-sop
.inria.fr/nachos/pics/results/hpc/Yguide-2.png
(:tableend:)
%center% Contour lines of the amplitude of the discrete Fourier transform of the electric field
to:
We conducted a parallel performance evaluation in terms of strong scalability analysis. For that purpose, we selected a use case typical of optical guiding applications. A Y-shaped waveguide is considered which consists in nanosphere embedded in vacuum. The constructed tetrahedral mesh consists of 520,704 vertices and 2,988,103 elements.  The high order discontinuous finite element method designed for the solution of the system of time-domain Maxwell equations coupled to a Drude model for the dispersion of noble metals at optical frequencies is formulated on a tetrahedral mesh. Within each element (tetrahedron) of the mesh, the components of the electric and magnetic field, as well as the component of the electric polarization, are approximated by a nodal (Lagrange type) interpolation method. The unknowns of the problem are thus given by the values of these physical quantities at the nodes of the polynomial interpolation. For instance, for a linear (i.e. P1) interpolation of the fields, the number of DoFs (Degrees of Freedoms) within a tetrahedron is 6x4 if the element is located in vacuum, and 9x4 if the element is located in the metallic structure. For a quadratic (i.e. P2) interpolation, the corresponding figures are 6x10 and 9x10, and so on for higher interpolation degrees. Then the global number of DoFs is the sum of these figures of the elements of the given mesh.
April 25, 2015, at 12:41 PM by 82.228.254.112 -
Changed lines 21-22 from:
(:cellnr align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-1.png
(:cell  align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
to:
(:cellnr align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-1.png
(:cell  align='center':) %width=350px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
April 25, 2015, at 12:41 PM by 82.228.254.112 -
Changed lines 21-22 from:
(:cellnr align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-1.jpg
(:cell  align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.jpg
to:
(:cellnr align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-1.png
(:cell  align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.png
Added line 24:
%center% Contour lines of the amplitude of the discrete Fourier transform of the electric field
April 25, 2015, at 12:38 PM by 82.228.254.112 -
Added lines 17-23:

(:linebreaks:)

(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-1.jpg
(:cell  align='center':) %width=250px% http://www-sop.inria.fr/nachos/pics/results/hpc/Yguide-2.jpg
(:tableend:)
April 25, 2015, at 12:37 PM by 82.228.254.112 -
Added line 16:
In this study, we focus on the last of these features and develop a hybrid coarse grain/fine grain parallelization of a high order DGTD solver formulated on unstructured tetrahedral meshes for the simulation of light interaction with nanometer scale metallic structures.
April 25, 2015, at 12:31 PM by 82.228.254.112 -
Deleted line 9:
Deleted line 10:
Deleted line 11:
Deleted line 12:
Deleted line 13:
April 25, 2015, at 12:30 PM by 82.228.254.112 -
Added lines 4-20:

This study has been conducted in the context of a PRACE Preparatory Access project (17th cut-off date, june 2014, project #2010PA2452).

(:linebreaks:)

The DGTD method can be considered as a finite element method where the continuity constraint at an element interface is released.  While it keeps almost all the advantages of the finite element method (large spectrum of applications, complex geometries, etc.), the DGTD method has other nice properties, which explain the renewed interest it gains in various domains in scientific computing:

* It is naturally adapted to a high order approximation of the unknown field.  Moreover, one may increase the degree of the approximation in the whole mesh as easily as for spectral methods but, with a DGTD method, this can also be done locally i.e. at the mesh cell level.  In most cases, the approximation relies on a polynomial interpolation method but the method also offers the flexibility of applying local approximation strategies that best fit to the intrinsic features of the modeled physical phenomena.

* When the discretization in space is coupled to an explicit time integration method, the DG method leads to a block diagonal mass matrix independently of the form of the local approximation (e.g the type of polynomial interpolation).  This is a striking difference with classical, continuous FETD formulations. Moreover, the mass matrix is diagonal if an orthogonal basis is chosen.

* It easily handles complex meshes.  The grid may be a classical conforming finite element mesh, a non-conforming one or even a hybrid mesh made of various elements  (tetrahedra, prisms, hexahedra, etc.).  The DGTD method has been proven to work well with highly locally refined meshes.  This property makes the DGTD method more suitable to the design of a hp-adaptive solution strategy  (i.e. where the characteristic mesh size h and the interpolation degree p changes locally wherever it is needed).

* It is flexible with regards to the choice of the time stepping scheme.  One may combine the discontinuous Galerkin spatial discretization with any global or local explicit time integration scheme, or even implicit, provided the resulting scheme is stable.

* It is naturally adapted to parallel computing.  As long as an explicit time integration scheme is used, the DGTD method is easily parallelized.  Moreover, the compact nature of method is in favor of high computation  to communication ratio  especially when the interpolation order is increased.

April 25, 2015, at 11:41 AM by 82.228.254.112 -
April 25, 2015, at 11:40 AM by 82.228.254.112 -
Changed lines 8-9 from:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.png
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.png
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.jpg
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p4.jpg
April 25, 2015, at 11:39 AM by 82.228.254.112 -
Changed lines 8-9 from:
(:cellnr align='center':) %width=320px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.png
(:cellnr align='center':) %width=320px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.png
to:
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.png
(:cellnr align='center':) %width=450px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.png
April 25, 2015, at 11:38 AM by 82.228.254.112 -
Added lines 4-10:

(:linebreaks:)

(:table border='0' width='100%' align='center' cellspacing='1px':)
(:cellnr align='center':) %width=320px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p2.png
(:cellnr align='center':) %width=320px% http://www-sop.inria.fr/nachos/pics/results/hpc/dgtd_p3.png
(:tableend:)
April 25, 2015, at 11:36 AM by 82.228.254.112 -
Added lines 1-5:
(:title Hybrid MIMD/SIMD high order DGTD solver:)

(:linebreaks:)

(:linebreaks:)