<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Dimension Reduction on large data sets in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Dimension-Reduction-on-large-data-sets/m-p/596393#M80006</link>
    <description>&lt;P&gt;I am often dealing with data that has thousands of terms, and many thousands, sometime hundreds of thousands, observations.&amp;nbsp; Many of the terms are usually correlated, so I would like to reduce the dimensionality.&amp;nbsp; But if I try to use Principal Components to do this the computation time is excessive or downright impractical.&amp;nbsp; Is there another method that would work better to reduce the dimensions of the data in such a case?&lt;/P&gt;</description>
    <pubDate>Thu, 08 Jun 2023 16:39:07 GMT</pubDate>
    <dc:creator>pcarroll1</dc:creator>
    <dc:date>2023-06-08T16:39:07Z</dc:date>
    <item>
      <title>Dimension Reduction on large data sets</title>
      <link>https://community.jmp.com/t5/Discussions/Dimension-Reduction-on-large-data-sets/m-p/596393#M80006</link>
      <description>&lt;P&gt;I am often dealing with data that has thousands of terms, and many thousands, sometime hundreds of thousands, observations.&amp;nbsp; Many of the terms are usually correlated, so I would like to reduce the dimensionality.&amp;nbsp; But if I try to use Principal Components to do this the computation time is excessive or downright impractical.&amp;nbsp; Is there another method that would work better to reduce the dimensions of the data in such a case?&lt;/P&gt;</description>
      <pubDate>Thu, 08 Jun 2023 16:39:07 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Dimension-Reduction-on-large-data-sets/m-p/596393#M80006</guid>
      <dc:creator>pcarroll1</dc:creator>
      <dc:date>2023-06-08T16:39:07Z</dc:date>
    </item>
    <item>
      <title>Re: Dimension Reduction on large data sets</title>
      <link>https://community.jmp.com/t5/Discussions/Dimension-Reduction-on-large-data-sets/m-p/596860#M80055</link>
      <description>&lt;P&gt;Some thoughts.&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;Are the data from the same source, e.g. they are similarly structured from data set to data set? If so, finding a more efficient way is worthwhile. Otherwise, you get one short deal every time anyway, and machine might still be faster each time.&lt;/LI&gt;
&lt;LI&gt;What is the next step after dimension reduction? What the objective? You many not need dimension reduction (or a separate dimension reduction) at all, if one methodology can address your objective without preprocessing to reduce dimension. Think of advanced capabilities in Generalized Regression.&lt;/LI&gt;
&lt;LI&gt;PCA operates on covariance matrix. So you have a large matrix and many rows to come up with that matrix. Would a heuristic dimension reduction method work, e.g. variable clustering?&lt;/LI&gt;
&lt;/OL&gt;</description>
      <pubDate>Fri, 03 Feb 2023 14:59:32 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Dimension-Reduction-on-large-data-sets/m-p/596860#M80055</guid>
      <dc:creator>peng_liu</dc:creator>
      <dc:date>2023-02-03T14:59:32Z</dc:date>
    </item>
  </channel>
</rss>

