KDD Competition 2010

From Noisebridge
(Difference between revisions)
Jump to: navigation, search
Line 54: Line 54:
* [http://swarmfinancial.com/screencasts/nb/kddWekaUsage1.swf Screencast1]
* [http://swarmfinancial.com/screencasts/nb/kddWekaUsage1.swf Screencast1]
* [http://swarmfinancial.com/screencasts/nb/kddWekaUsage2.swf Screencast2]
* [http://swarmfinancial.com/screencasts/nb/kddWekaUsage2.swf Screencast2]
== How to run SVM ==
* See the notes at [[Machine Learning/SVM]]

Revision as of 22:46, 21 May 2010

We're interested in working on the KDD Competition, as a way to focus our machine learning exploration -- and maybe even finding some interesting aspects to the data! If you're interested, drop us a note, show up at a weekly Machine Learning meeting, and we'll use this space to keep track of our ideas.




  • Vikram -- will help setting up Hadoop for the rest of us & create a guide for Mahout setup
  • Thomas -- will get libsvm working on the data and put together a "how to" guide for doing so
    • put together a perl script which will take random samples from the data, for working on smaller instances
    • put together a simple R script for loading the data
  • Andy -- will get Weka working on the data and put together a "how to" guide for doing so
  • Erin -- Will put meeting notes of 5/19 on https://www.noisebridge.net/wiki/Machine_Learning; will work on data transformations and ways to create better representations of the data; will provide the orthogonalized data sets
  • We will need to make sure we don't get disqualified for people belonging to multiple teams! Do not sign up anybody else for the competition without asking first.


  • For KDD submission: to zip the submission file on OSX: use command line, otherwise will complain about __MACOSX file: e.g.: zip asdf.zip algebra_2008_2009_submission.txt


  • Add new features by computing their values from existing columns -- e.g. correlation between skills based on their co-occurence within problems. Could use Decision tree to define boundaries between e.g. new "good student, medium student, bad student" feature
  • Dimensionality reduction -- transform into numerical values appropriate for consumption by SVM

Who we are

  • Andy; Machine Learning
  • Thomas; Statistics
  • Erin; Maths
  • Vikram; Hadoop

(insert your name/contact info/expertise here)

How to run Weka (quick 'n dirty tutorial)

  • Download and install Weka
  • Get your KDD data
  • preprocess your data: this command takes 1000 lines from the given training data set and converts it into .csv file
  • attention, in the last sed command you need to replace the long whitespace with a tab. In OSX terminal, you do that by pressing CONTROL+V and then tab. (Copying and pasting the command below won't work, since it interprets the whitespace as spaces)
  • head -n 1000 algebra_2006_2007_train.txt | sed -e 's/[",]/ /g' | sed 's/ /,/g' > algebra_2006_2007_train_1kFormatted.csv
  • The following screencast shows you how to do these steps:
  • In Weka's Explorer, remove some unwanted attributes (I leave this up to your judgment), inspect the dataset.
  • Then you can run a ML algorithm over it, e.g. Neural Networks to predict the student performance.
  • Screencast1
  • Screencast2

How to run SVM

Personal tools