Adaptive Support Vector Machines (aSVM) is a software package implementing an algorithm for adapting one or multiple existing SVM (source) classifiers to a SVM (target) classifier for a new domain. For example, if there are two data domains as A and B with similar distributions, one can use aSVM to adapt a classifier on A to a classifier for B based on a small number of labeled data in B. The adapted classifier is expected perform better than a classifier trained only from the same amount of labeled examples. This is implemented based on the LIBSVM package.
The goal of aSVM is to improve classification on a new domain by leveraging the existing models from related domains. It has many real-world applications. For example, if you have a SVM model for classifying indoor/outdoor images trained from Corel collection, you can adapt it to classify indoor/outdoor photos in Flickr collection by labeling only a small number of photos. This paper contains some results of using this algorithm for classifying video scenes from different sources.
This package has the following features:
- adapt one or multiple source models
- average, manual, or automatic weighting of source model(s)
- accept standard data format (same as LIBSVM)
- output standard SVM model file (same as LIBSVM)
the executable (only for Windows)
the source code in C++ (as Visual C++ project)
adapt_svm_train [options] [+ source_model_file[,weight]] train_data_file [output_model_file]
-t (default 2): Type of kernel function; 0 for linear, 1 for polynomial, 2 for radial basis function (RBF), 3 for sigmoid
-d (default 3): Degree in polynomial kernel function
-g (default 1/k, k as feature dimension): Gamma parameter in RBF kernel function
-c (default 1): Cost parameter C on classification error
-m (default 100): Cache memory size in MB
-e (default 0.001): Tolerance (epsilon in SVM) of termination criterion
-h (default 1): whether to use shrinking heuristics, 0 or 1
-wi (default 1): set the parameter C of class i to weight*C (e.g., -w1 1 –w2 1)
-l (default 0): whether to learn the weights of source models, 0 or 1
-s (default 1): Contribution factor of source models, effective only when –l 1
The training data file contains the labeled examples from the target domain. The file must follow the format of LIBSVM’s data file, which looks like:
<label>,<ID>:< value> ...
<label>,<ID>:<value> ... <ID>:<value>
Each row represents one data point, where <label> is the label (-1 or 1), and <ID>:<value> pair denotes one of its feature components, with <feature> as the ID of the feature (i.e., order starting from 1). Zero-valued features can be omitted. Thus, each line can be of different length, depending on the number of non-zero features for each data point.
An example data file is given below, where each data point has a 10-d feature vector.
1 1:0.4 6:0.1 10:0.7
-1 3:0.2 7:0.15
-1 1:0.8 2:0.7 8:0.2 9:0.1
-1 2:0.1 4:0.3 7:0.7
adapt_svm_train –t 2 –g 0.1 + model1.txt,0.8 + model2.txt,0.2 train.dat
adapt_svm_train –t 0 –l 1 –s 100 + model1.txt + model2.txt train.dat model.txt
If you find bugs or you have problems with the code you cannot solve by yourself, please contact me via email email@example.com.
This software is free only for non-commercial use. It must not be modified and distributed without prior permission of the author. The author is not responsible for implications from the use of this software.
Jun Yang, Rong Yan, Alex Hauptmann, Cross-Domain Video Concept Detection using Adaptive SVMs, ACM Multimedia 2007.
* Last modified May 26th, 2008 by Jun Yang