Machine Learning: Difference between revisions
Mschachter (talk | contribs) |
No edit summary |
||
(29 intermediate revisions by 7 users not shown) | |||
Line 1: | Line 1: | ||
=== Next Meeting=== | === Next Meeting=== | ||
*When: | *When: | ||
*Where: 2169 Mission St. (back corner, Church | *Where: 2169 Mission St. (back NE corner, Church classroom) | ||
*Topic: | *Topic: | ||
*Details: | *Details: Currently on hiatus until somebody decides to pick it back up! | ||
*Who: | *Who: | ||
=== Take the Noisebridge ML Survey === | === Take the Noisebridge ML Survey === | ||
Line 14: | Line 14: | ||
=== About Us === | === About Us === | ||
We're a loosely-knit stochastic federation of people who like Noisebridge and like machine learning. What is machine learning? It's broad field that typically involves training computer models to solve problems. How can you participate? Join the [https://www.noisebridge.net/mailman/listinfo/ml mailing list], send an email and introduce yourself. Show up to the next meeting, share your thoughts. Participate in projects or start your own. Go to workshops, write code at workshops, learn stuff, give workshops of your own! All are welcome. | We're a loosely-knit stochastic federation of people who like Noisebridge and like machine learning. What is machine learning? It's broad field that typically involves training computer models to solve problems. How can you <span class="plainlinks">[http://www.monoloop.com<span style="color:black;font-weight:normal; text-decoration:none!important; background:none!important; text-decoration:none;">website personalization</span>] participate? Join the [https://www.noisebridge.net/mailman/listinfo/ml mailing list], send an email and introduce yourself. Show up to the next meeting, share your thoughts. Participate in projects or start your own. Go to workshops, write code at workshops, learn stuff, give workshops of your own! All are welcome. | ||
=== Talks and Workshops === | === Talks and Workshops === | ||
Line 34: | Line 34: | ||
=== Future Talks and Topics, Ideas === | === Future Talks and Topics, Ideas === | ||
*Random Forests in R | *Random Forests in R | ||
*Restricted Boltzmann Machines (Mike S, some day) | *Restricted Boltzmann Machines (Mike S, some day) | ||
*Analyzing brain cells (Mike S) | *Analyzing brain cells (Mike S) | ||
Line 69: | Line 69: | ||
*[http://www.ntis.gov/products/ssa-dmf.aspx Social Security Death Master File!] | *[http://www.ntis.gov/products/ssa-dmf.aspx Social Security Death Master File!] | ||
=== Software Tools | === Software Tools === | ||
==== Generic ML Libraries ==== | ==== Generic ML Libraries ==== | ||
Line 96: | Line 96: | ||
*[http://nipy.sourceforge.net/nitime/ nitime] | *[http://nipy.sourceforge.net/nitime/ nitime] | ||
**NeuroImaging in Python, has some good time series analysis stuff and multi-variate response fitting. | **NeuroImaging in Python, has some good time series analysis stuff and multi-variate response fitting. | ||
*[http://pandas.pydata.org/ Pandas] | |||
**Data analysis workflow in python | |||
*[http://www.pytables.org/moin PyTables] | |||
**Adds querying capabilities to HDF5 files | |||
*[http://statsmodels.sourceforge.net/ statsmodels] | |||
**Regression, time series analysis, statistics stuff for python | |||
*[https://github.com/JohnLangford/vowpal_wabbit/wiki Vowpal Wabbit] | |||
**"Intrinsically Fast" implementation of gradient descent for large datasets | |||
*[http://www.shogun-toolbox.org/ Shogun] | |||
**Fast implementations of SVMs | |||
*[http://mc-stan.org/ Stan] | |||
**A graphical model compiler | |||
==== Computer Vision ==== | ==== Computer Vision ==== | ||
Line 108: | Line 120: | ||
*[http://code.google.com/p/pyo/ pyo] | *[http://code.google.com/p/pyo/ pyo] | ||
**Real-time audio signal processing | **Real-time audio signal processing | ||
*[https://github.com/jsawruk/pymir PYMir] | |||
**A library for reading mp3's into python, and doing analysis | |||
*[http://wiki.python.org/moin/PythonInMusic List of Sound Tools for Python] | |||
==== Data Visualization ==== | ==== Data Visualization ==== | ||
Line 180: | Line 195: | ||
=== [[Machine Learning/Meeting Notes|Meeting Notes]]=== | === [[Machine Learning/Meeting Notes|Meeting Notes]]=== | ||
[[Category:Events]] | |||
[[Category:Projects]] |
Revision as of 23:37, 5 February 2013
Next Meeting
- When:
- Where: 2169 Mission St. (back NE corner, Church classroom)
- Topic:
- Details: Currently on hiatus until somebody decides to pick it back up!
- Who:
Take the Noisebridge ML Survey
Take a survey and vote for what you want to learn!
Crowdsourced Q&A
Are you working on a data mining, machine learning, or statistics problem? Do you want some help? Consider sending an email to the mailing list about it! Also consider setting up a day to come in and talk about the project you're working on and get input from Spotsylvania reckless driving other ML people.
About Us
We're a loosely-knit stochastic federation of people who like Noisebridge and like machine learning. What is machine learning? It's broad field that typically involves training computer models to solve problems. How can you website personalization participate? Join the mailing list, send an email and introduce yourself. Show up to the next meeting, share your thoughts. Participate in projects or start your own. Go to workshops, write code at workshops, learn stuff, give workshops of your own! All are welcome.
Talks and Workshops
We've given lots of workshops and talks over the past year or so, here's a few. Many of the workshops we've given previously are recurring and will be given again, especially upon request!
- Intro to Machine Learning
- A Brief Tour of Statistics
- Generalized Linear Models
- Neural Nets Workshop
- Support Vector Machines
- Random Forests
- Independent Components Analysis
- Deep Nets
Code and SourceForge Site
- We have a Sourceforge Project
- We have a git repository on the project page, accessible as:
git clone git://ml-noisebridge.git.sourceforge.net/gitroot/ml-noisebridge/ml-noisebridge
- Send an email to the list if you want to become an administrator on the site to get write access to the git repo!
Future Talks and Topics, Ideas
- Random Forests in R
- Restricted Boltzmann Machines (Mike S, some day)
- Analyzing brain cells (Mike S)
- Deep Nets w/ Stacked Autoencoders (Mike S, some day)
- Generalized Linear Models (Mike S, Erin L? some day)
- Graphical Models
- Working with the Kinect
- Computer Vision with OpenCV
Mailing List
https://www.noisebridge.net/mailman/listinfo/ml
Projects
- Small Group Subproblems
- Fundraising
- Noisebridge Machine Learning Course
- Kaggle Social Network Contest
- KDD Competition 2010
- HIV
Datasets and Websites
- UCI Machine Learning Repository
- DataSF.org
- Infochimps
- Face Recognition Databases
- Time Series Data Library
- Data Q&A Forum
- Metaoptimize
- Quora ML Page
- A ton of Weather Data
- MLcomp
- Upload your algorithm and objectively compare it's performance to other algorithms
- Social Security Death Master File!
Software Tools
Generic ML Libraries
- Weka
- a collection of data mining tools and machine learning algorithms.
- MOA (Massive Online Analysis)
- Offshoot of weka, has all online-algorithms
- scikits.learn
- Machine learning Python package
- scikits.statsmodels
- Statistical models to go with scipy
- PyBrain
- Does feedforward, recurrent, SOM, deep belief nets.
- LIBSVM
- c-based SVM package
- PyML
- MDP
- Modular framework, has lots of stuff!
- VirtualBox Virtual Box Image with Pre-installed Libraries listed here
- Theano: Symbolic Expressions and Transparent GPU Integration
- sympy Does symbolic math
- Waffles
- Open source C++ set of machine learning command line tools.
- RapidMiner
- Mobile Robotic Programming Toolkit
- nitime
- NeuroImaging in Python, has some good time series analysis stuff and multi-variate response fitting.
- Pandas
- Data analysis workflow in python
- PyTables
- Adds querying capabilities to HDF5 files
- statsmodels
- Regression, time series analysis, statistics stuff for python
- Vowpal Wabbit
- "Intrinsically Fast" implementation of gradient descent for large datasets
- Shogun
- Fast implementations of SVMs
- Stan
- A graphical model compiler
Computer Vision
Audio Processing
- Friture
- Real-time spectrogram generation
- pyo
- Real-time audio signal processing
- PYMir
- A library for reading mp3's into python, and doing analysis
- List of Sound Tools for Python
Data Visualization
- Orange
- Strong data visualization component
- Gephi
- Graph Visualization
- ggplot
- Nice plotting package for R
- MayaVi2
- 3D Scientific Data Visualization
Cluster Computing
- Mahout
- Hadoop cluster based ML package.
- STAR: Cluster
- Easily build your own Python computing cluster on Amazon EC2
Other
Presentations and other Materials
- Awesome Machine Learning Applications -- A list of cool applications of ML
- Hands-on Machine Learning, a presentation jbm gave on 2009-01-07.
- http://www.youtube.com/user/StanfordUniversity#g/c/A89DCFA6ADACE599 Stanford Machine Learning online course videos]
- Media:Brief_statistics_slides.pdf, a presentation given on statistics for the machine learning group
- LinkedIn discussion on good resources for data mining and predictive analytics
- Face Recognition Algorithms
- Max Welling's ML classnotes
Topics to Learn and Teach
NBML Course - Noisebridge Machine Learning Curriculum (work-in-progress)
CS229 - The Stanford Machine learning Course @ noisebridge
- Supervised Learning
- Linear Regression
- Linear Discriminants
- Neural Nets/Radial Basis Functions
- Support Vector Machines
- Classifier Combination [1]
- A basic decision tree builder, recursive and using entropy metrics
- Unsupervised Learning
- Hidden Markov Models
- Clustering: PCA, k-Means, Expectation-Maximization
- Graphical Modeling
- Generative Models: gaussian distribution, multinomial distributions, HMMs, Naive Bayes
- Deep Belief Networks & Restricted Boltzmann Machines
- Reinforcement Learning
- Temporal Difference Learning
- Math, Probability & Statistics
- Metric spaces and what they mean
- Fundamentals of probabilities
- Decision Theory (Bayesian)
- Maximum Likelihood
- Bias/Variance Tradeoff, VC Dimension
- Bagging, Bootstrap, Jacknife [2]
- Information Theory: Entropy, Mutual Information, Gaussian Channels
- Estimation of Misclassification [3]
- No-Free Lunch Theorem [4]
- Machine Learning SDK's
- Applications
- Collective Intelligence & Recommendation Engines