<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
	<channel>
		<atom:link href="http://sobereva.com/wfnbbs/extern.php?action=feed&amp;tid=752&amp;type=rss" rel="self" type="application/rss+xml" />
		<title><![CDATA[Multiwfn forum / Patch: omp collapse(2) in grid.f90]]></title>
		<link>http://sobereva.com/wfnbbs/viewtopic.php?id=752</link>
		<description><![CDATA[The most recent posts in Patch: omp collapse(2) in grid.f90.]]></description>
		<lastBuildDate>Sun, 18 Dec 2022 09:38:33 +0000</lastBuildDate>
		<generator>FluxBB</generator>
		<item>
			<title><![CDATA[Re: Patch: omp collapse(2) in grid.f90]]></title>
			<link>http://sobereva.com/wfnbbs/viewtopic.php?pid=2901#p2901</link>
			<description><![CDATA[<p>Dear Igor,</p><p>collapse(2) is really fantastic! Your patch has been merged into official source code.</p><p>I tested 704atoms.wfn on my dual AMD EPYC 7R32 (96 physical cores) server, the costs using new version for calculating high quality grid data of electron density and ELF are 2s and 6s, respectively. While the costs using old version are 5s and 20s. The speed-up by collapse(2) on the server with large number of cores is surprisingly high!</p><p>However, I removed &quot;if(mod(ifinish,256)==0)&quot;, otherwise after calculation I will observe</p><div class="codebox"><pre><code>Calculation of grid data took up wall clock time         2 s-]   99.89 %     /</code></pre></div><p>Namely the progress bar is not 100%. My brief test showed that removing &quot;if(mod(ifinish,256)==0)&quot; doesn&#039;t detectably hurt performance, at least on my 8-core notebook and 96-core server.</p><p>Best regards,</p><p>Tian</p>]]></description>
			<author><![CDATA[dummy@example.com (sobereva)]]></author>
			<pubDate>Sun, 18 Dec 2022 09:38:33 +0000</pubDate>
			<guid>http://sobereva.com/wfnbbs/viewtopic.php?pid=2901#p2901</guid>
		</item>
		<item>
			<title><![CDATA[Re: Patch: omp collapse(2) in grid.f90]]></title>
			<link>http://sobereva.com/wfnbbs/viewtopic.php?pid=2891#p2891</link>
			<description><![CDATA[<p>Dear Igor,</p><p>Thanks, I&#039;ll check and test shortly. I just infected with COVID-19 and my productivity has been greatly affected, so it may take longer time for me to give you reponse...</p><p>Best regards,</p><p>Tian</p>]]></description>
			<author><![CDATA[dummy@example.com (sobereva)]]></author>
			<pubDate>Fri, 16 Dec 2022 06:02:06 +0000</pubDate>
			<guid>http://sobereva.com/wfnbbs/viewtopic.php?pid=2891#p2891</guid>
		</item>
		<item>
			<title><![CDATA[Patch: omp collapse(2) in grid.f90]]></title>
			<link>http://sobereva.com/wfnbbs/viewtopic.php?pid=2886#p2886</link>
			<description><![CDATA[<p>Dear Tian,</p><p>As I mentioned on <a href="http://sobereva.com/wfnbbs/viewtopic.php?id=732" rel="nofollow">http://sobereva.com/wfnbbs/viewtopic.php?id=732</a> topic, I found a way to have a speed-up.</p><p>Here, the patch is presented. The patch is affected for machines with a large number of threads. Probably, a similar patch can be applied through the whole code.</p><p><a href="http://sobereva.com/wfnbbs/uploads/267c6d3e9e9b48478a545c8726ac2605.txt" rel="nofollow">Multiwfn_collapse.patch.txt</a></p><p>The effect of the patch I tested on 704atoms.wfn. Here, the speed-ups are presented for a different number of cores. The black line means ideal scale. After the patch, the ideal scale is up to 26 cores, while before only up to 19 (?).</p><p><a href="https://postimg.cc/R60hC3tF" rel="nofollow"><span class="postimg"><img src="https://i.postimg.cc/R60hC3tF/collapse.png" alt="collapse.png" /></span></a></p><p>Probably, for better scalability, I need a larger system (or a slower computer) since even for code without collapse, near 32 cores, time became about 5 seconds, and for `collapse(2)`, time became about 3 seconds for 32 cores.</p><p>Best regards,<br />Igor</p>]]></description>
			<author><![CDATA[dummy@example.com (i.s.ger)]]></author>
			<pubDate>Thu, 15 Dec 2022 13:31:45 +0000</pubDate>
			<guid>http://sobereva.com/wfnbbs/viewtopic.php?pid=2886#p2886</guid>
		</item>
	</channel>
</rss>
