开发者

training for classification using libsvm

开发者 https://www.devze.com 2022-12-22 06:13 出处:网络
I want to classify using libsvm. I have 9 training sets , each set has 144000 labelled instances , each instance having a variable number of features. It is taking about 12 hours to train one set ( ./

I want to classify using libsvm. I have 9 training sets , each set has 144000 labelled instances , each instance having a variable number of features. It is taking about 12 hours to train one set ( ./svm-train with probability estimates ). As i dont have much time , I would like to run more than one set at a time. I'm not sure if i can do this.. Can i run all 9 processes simultaneously in different terminals ?

./svm-train -b 1 feat1.txt
./svm-tra开发者_运维技巧in -b 1 feat2.txt
      .
      .
      .
./svm-train -b 1 feat9.txt

( i'm using fedora core 5 )


You can tell libsvm to use openmp for parallelization. Look at this libsvm faq entry: http://www.csie.ntu.edu.tw/~cjlin/libsvm/faq.html#f432


As Adam said, it depends on how many cores and processors your system has available. If that's insufficient, why not spin up a few EC2 instances to run on?

The Infochimps MachetEC2 public AMI comes with most of the tools you'll need: http://blog.infochimps.org/2009/02/06/start-hacking-machetec2-released/


Yes. But unless you have a multi-core or multi-processor system it may not save you that much time.

0

精彩评论

暂无评论...
验证码 换一张
取 消