<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Feature selection in JMP in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34911#M20620</link>
    <description>&lt;P&gt;Hi&lt;/P&gt;&lt;P&gt;I knwow that in JMP Pro for feature selection (I'm not interested in future extraction...) there is the Generalized regression module. Do you know if&amp;nbsp;a similar module is in JMP (standard version) too? If not&amp;nbsp;which is the best method to use with JMP12/13 to reduce the number of predictors to&amp;nbsp;leave only the most important ones?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&amp;nbsp; Felice&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Tue, 24 Jan 2017 16:35:00 GMT</pubDate>
    <dc:creator>FR60</dc:creator>
    <dc:date>2017-01-24T16:35:00Z</dc:date>
    <item>
      <title>Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34911#M20620</link>
      <description>&lt;P&gt;Hi&lt;/P&gt;&lt;P&gt;I knwow that in JMP Pro for feature selection (I'm not interested in future extraction...) there is the Generalized regression module. Do you know if&amp;nbsp;a similar module is in JMP (standard version) too? If not&amp;nbsp;which is the best method to use with JMP12/13 to reduce the number of predictors to&amp;nbsp;leave only the most important ones?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&amp;nbsp; Felice&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 24 Jan 2017 16:35:00 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34911#M20620</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2017-01-24T16:35:00Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34920#M20623</link>
      <description>&lt;P&gt;Hello Felice,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;There is no Generalized Regression per se in regular JMP. &amp;nbsp;You do have a few&amp;nbsp;of options though. &amp;nbsp;If you have highly correlated factors/inputs you can use either Partial Least Squares or Principal Component Analysis. &amp;nbsp;One other option would be to use Partition Analysis/Decision Trees. &amp;nbsp;Be sure to use a holdout set to check the predictive capability of any model you build. &amp;nbsp;For partition analysis look under the red hot spot at the top and select Column Contributions to see which factors are most important.&lt;/P&gt;
&lt;P&gt;Hope this helps.&lt;/P&gt;
&lt;P&gt;Bill&lt;/P&gt;</description>
      <pubDate>Tue, 24 Jan 2017 17:42:04 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34920#M20623</guid>
      <dc:creator>Bill_Worley</dc:creator>
      <dc:date>2017-01-24T17:42:04Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34921#M20624</link>
      <description>Hi Felice,&lt;BR /&gt;&lt;BR /&gt;There is something called Predictor Screening as well in JMP. Here is a link about that option.&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://www.jmp.com/support/help/13/Predictor_Screening.shtml" target="_blank"&gt;http://www.jmp.com/support/help/13/Predictor_Screening.shtml&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;That can help reduce the number of possible predictors and is in JMP.&lt;BR /&gt;</description>
      <pubDate>Tue, 24 Jan 2017 17:48:07 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34921#M20624</guid>
      <dc:creator>Chris_Kirchberg</dc:creator>
      <dc:date>2017-01-24T17:48:07Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34926#M20626</link>
      <description>&lt;P&gt;You've gotten some good choices from the previous responses (some of which I will repeat), but there are others.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You could just use stepwise regression (a classic). You could use All-Subsets regression. You could use variable clustering. You could use Partition. You could use PLS. You could use PCR (principal component regression). There are graph-based methods: normal plots, pareto plot, bayes plot. You could use predictor screening, etc.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;There is a new course being created right now called "JMP Software: Finding Important Predictors" that addresses this issue and covers these techniques, using both JMP and JMP Pro. This two day class will likely be available for delivery at customer locations starting in April. It will be offered as a public class for the first time at the Discovery conference this October in St. Louis.&lt;/P&gt;</description>
      <pubDate>Tue, 24 Jan 2017 20:16:05 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34926#M20626</guid>
      <dc:creator>Dan_Obermiller</dc:creator>
      <dc:date>2017-01-24T20:16:05Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34937#M20632</link>
      <description>&lt;P&gt;In addition to all the great suggestions from my colleagues above, if you have a relatively small number of variables I'd also look at the Multivariate Methods -&amp;gt; Multivariate -&amp;gt; Scatterplot Matrix platform. This will give you a nice matrix of pairwise correlations that may exist among the variables and this can be valuable in trying to discover correlations that might exist within the predictor set. These correlations don't play nice with some regression procedures such as ordinary least squares regression (called Standard Least Squares in JMP). If you go down the standard least squares path, make sure to check the variance inflation factors in the parameter estimates table for indications of predictor correlations that may be present.&lt;/P&gt;</description>
      <pubDate>Wed, 25 Jan 2017 12:03:26 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34937#M20632</guid>
      <dc:creator>Peter_Bartell</dc:creator>
      <dc:date>2017-01-25T12:03:26Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34948#M20641</link>
      <description>&lt;P&gt;Great&lt;/P&gt;&lt;P&gt;I used it. Very nice.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Felice&lt;/P&gt;</description>
      <pubDate>Wed, 25 Jan 2017 15:46:00 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34948#M20641</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2017-01-25T15:46:00Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34951#M20642</link>
      <description>&lt;P&gt;&lt;FONT color="#ff0000"&gt;Ciao Dan&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;thank you very much for your msg.&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;I have some comment to do....&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You could just use stepwise regression (a classic).&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;Can it manipulate more than 1K predictors?&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You could use variable clustering.&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;Can you give me more details on this tecnique on how to choose&amp;nbsp;important predictors through clustering? &lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You could use Partition.&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;Tipically in our Fab we&amp;nbsp;use it&amp;nbsp;after removing not important predictors (let's consider that&amp;nbsp;generally we have more than thousands predictors and a lot of noisy in our data ....)&amp;nbsp;&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You could use PCR (principal component regression).&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;We know this but we loose information on predictor meaning and then we prefer don't use it.&amp;nbsp;&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;There are graph-based methods: normal plots, pareto plot, bayes plot.&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;I don't know how to do feature selection with graph. Sorry.&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;There is a new course being created right now called "JMP Software: Finding Important Predictors" that addresses this issue and covers these techniques, using both JMP and JMP Pro. This two day class will likely be available for delivery at customer locations starting in April.&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;This is a great news. I hope that will be available for Italy too. If yes for sure I will follow them.&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;Rgds.&amp;nbsp; Felice&lt;/FONT&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 25 Jan 2017 16:06:36 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34951#M20642</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2017-01-25T16:06:36Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34952#M20643</link>
      <description>&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;
&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;My answers to your questions are in blue.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/7746"&gt;@FR60&lt;/a&gt; wrote:&lt;BR /&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;Ciao Dan&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;thank you very much for your msg.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;I have some comment to do....&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You could just use stepwise regression (a classic).&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;Can it manipulate more than 1K predictors?&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;Yes. If you have the memory on your machine to handle large problems. I just ran a simple example with 10,000 observations and 2,000 predictors. Stepwise worked fine.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You could use variable clustering.&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;Can you give me more details on this tecnique on how to choose&amp;nbsp;important predictors through clustering? &lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;This would be an approach that is similar to principal components analysis, but instead of you looking at loading plots to see similar variables, JMP will cluster them for you automatically. You can then choose the variable that is most representative of the cluster or even create the "typical" variable for the cluster. This will help you avoid the "redundant information" you often see with many variables.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You could use Partition.&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;Tipically in our Fab we&amp;nbsp;use it&amp;nbsp;after removing not important predictors (let's consider that&amp;nbsp;generally we have more than thousands predictors and a lot of noisy in our data ....)&amp;nbsp;&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;There are many ways to use Partition, but think of that very first split. The approach needs to determine which split contains the most information. That is a variable selection. You could also use a trick that Dick DeVeaux calls "shaking the tree". Split many, many, times then look at the column contributions of the variables to identify the most important ones. So many ways to use this flexible platform!&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You could use PCR (principal component regression).&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;We know this but we loose information on predictor meaning and then we prefer don't use it.&amp;nbsp;&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;Understood. I am not a big fan of PCR for this reason. However, by looking at the loadings of the variables for only the significant principal components, you could possibly identify the original variables that are important. Use those important original variables to start building your model. There is nothing that says you must stick with the principal components.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;There are graph-based methods: normal plots, pareto plot, bayes plot.&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;I don't know how to do feature selection with graph. Sorry.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;Nothing to be sorry about. That is why we are creating the class. People are often unaware of these tools.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;There is a new course being created right now called "JMP Software: Finding Important Predictors" that addresses this issue and covers these techniques, using both JMP and JMP Pro. This two day class will likely be available for delivery at customer locations starting in April.&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;This is a great news. I hope that will be available for Italy too. If yes for sure I will follow them.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;All of the classes that we create at SAS in the U.S. are available to the international SAS offices, too. If they do not have an instructor that knows the topic, they can request one from another region that does have the skill set. Just ask your local SAS office for the training!&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#ff0000"&gt;Rgds.&amp;nbsp; Felice&lt;/FONT&gt;&lt;/P&gt;
&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;I hope my comments have been helpful.&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&lt;FONT color="#0000FF"&gt;Dan&lt;/FONT&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 25 Jan 2017 16:49:13 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34952#M20643</guid>
      <dc:creator>Dan_Obermiller</dc:creator>
      <dc:date>2017-01-25T16:49:13Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34953#M20644</link>
      <description>&lt;P&gt;&lt;FONT color="#ff0000"&gt;Ciao Dan&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;your comments were really welcome and more than helpful. Just last question about clustering.&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#0000ff"&gt;This would be an approach that is similar to principal components analysis, but instead of you looking at loading plots to see similar variables, JMP will cluster them for you automatically. You can then choose the variable that is most representative of the cluster or even create the "typical" variable for the cluster. This will help you avoid the "redundant information" you often see with many variables.&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;Once I have n &amp;nbsp;cluster&amp;nbsp;(some rule&amp;nbsp;on how choose the size of this&amp;nbsp;number ..)&amp;nbsp;how I can identify&amp;nbsp;the most representative variable for each cluster?&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;FONT color="#ff0000"&gt;Thanks.&amp;nbsp; Felice&amp;nbsp;&lt;/FONT&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 25 Jan 2017 17:09:35 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34953#M20644</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2017-01-25T17:09:35Z</dc:date>
    </item>
    <item>
      <title>Re: Feature selection in JMP</title>
      <link>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34954#M20645</link>
      <description>&lt;P&gt;JMP will tell you which variable is most representative of the cluster. It will also determine the proper number of clusters to use.&lt;/P&gt;</description>
      <pubDate>Wed, 25 Jan 2017 17:31:19 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Feature-selection-in-JMP/m-p/34954#M20645</guid>
      <dc:creator>Dan_Obermiller</dc:creator>
      <dc:date>2017-01-25T17:31:19Z</dc:date>
    </item>
  </channel>
</rss>

