Author Title Author Title Author Item Recommendation for ...clu/doc/ijcai16_AmpCMF_poster.pdf · m...

1
Item Recommendation for Emerging Online Businesses Chun-Ta Lu ([email protected]), Sihong Xie, Weixiang Shao, Lifang He and Philip S. Yu 1.Item Recommendation for Emerging Businesses (I) Emerging Business: new network with very sparse information. (II) Partially Aligned Networks: networks sharing common entities crowd-sourced review site ? ? follow review aligned entity 5 4 4 5 5 3 ? 3 4 C 1 C 2 C 3 C 4 P 1 P 2 P 3 U 1 U 2 U 3 U 4 U 5 I 1 I 2 I 3 I 4 4 ? electronic art sport rate tag E-commerce site How to measure similarity between users/items with sparse information? How to utilize information from aligned networks to improve performance? Hint: Transfer knowledge through aligned entities. P 5.Acknowledgments This work is supported in part by NSF through grants III-1526499, and CNS-1115234, NSFC through grant 61503253, and Google Research Award. 2. Augmented Meta Path-based Similarity (AmpSim) Product (P) Customer (C) User (U) Item (I) Rate Review Score Star Follow Partially aligned Partially aligned Tag (T) Has Schema of partially aligned networks Examples of Augmented Meta Path (AMP) Normalize link attributes to deal with different types of link attributes upon an AMP : Average over multiple AMPs: How to measure similarity via AMPs? Intra-network augmented meta path (C [star ] ---! P [star ] --- C ),(P [star ] --- C [star ] ---! P ), (P ! T P ) (C [star ] ---! P [star ] --- C [star ] ---! P [star ] --- C ) (P [star ] --- C [star ] ---! P [star ] --- C [star ] ---! P ) Inter-network augmented meta path (P ! I [score] ---- U [score] ----! I ! P ) (C ! U [score] ----! I [score] ---- U ! C ), (C ! U ! U ! C ) both linkage structures and augmented link attributes are taken into account Q l i=1 M i tes upon P is the product of the normalized link attributes upon Ordinal attributes: b i = Y i Other attributes: b i =0 MM T is (adjusted) cosine similarity After normalization, AmpSim: AMP-based similarity measure between v a based on path Y : link attributes M ui = Y ui - b i q P j (Y uj - b j ) 2 s(v a ,v b |P )= [ Q l i=1 M i ] ab +[ Q 1 i=l M T i ] ba [ Q l i=1 M i ] a+[ Q 1 i=l M T i ] b2 [0, 1] S (v a ,v b )= tanh( P i w i S i (v a ,v b )) tanh(1) 2 [0, 1] path direction is taken into account P i w i =1 v b and P 3. Recommendation Model: AmpCMF Collective Matrix Factorization (CMF) w/ alignment regularization geometric regularization min P () ,Q () 0, 2{s,t} J = X 2{s,t} ||I () (Y () - P () Q () T )|| 2 F + β (R A (P (t) , P (s) )+ R A (Q (t) , Q (s) )) + λ(R G (P (t) )+ R G (Q (t) )) X + 2 X 2{s,t} (||P () || 2 F + ||Q () || 2 F ) prevent overfitting alignment regularization: geometric regularization: Solve the optimization problem using multiplicative updates Y and I 2 R nm are user-item feedback and indicator matrices P 2 R nk and Q 2 R mk are low rank representations of users and items R G (P)= 1 2 X u,v S uv ||p u - p v || 2 2 = 1 2 Tr (P T (D - S)P)= 1 2 Tr (P T L P P) R A (P (t) , P (s) )= 1 2 X (i,g i )2A (t,s) ||p (t) i - p (s) g i || 2 2 = 1 2 kAA T P (t) - AP (s) k 2 F Partial Split 4. Experiment Results cold-start: test on users who have 5 ratings in training set 90% (55%) of users are cold-start users in Yelp (Epinions) Yelp: test a mature business entering a new domain target network (20% of ratings) source network (80% of ratings) overlapping (20% of users; all items) target network: Nevada w/ 11.5K users; 2.1K items source network: Arizona w/ 19.5K users; 6.3K items overlapping: 4.3K users; no items WNMF: weighted nonnegative matrix factorization Hete-MF: WNMF+ PathSim between items only Hete-CF: WNMF + PathSim between users/items Hete-PRW: WNMF + Pairwise RW similarity between users/items Amp-MF: WNMF + AmpSim between users/items CMF: collective matrix factorization w/ alignment regularization RMGM: rating-matrix generative model Amp-CMF: proposed method WNMF + geometric regularization Transfer learning based methods Epinions (Overall ) Epinions (Cold Start) Yelp (Overall ) Yelp (Cold Start) Epinions: test an emerging business partially aligned with a developed business Conclusion: 1. Methods using more accurate similarity measure achieve better performance 2. Combining both geometric and alignment regularization boosts the performance 3. Learning from highly overlapped networks can achieve better improvement: improve 21% in Yelp and 29% in Epinions Method RMSE MAE RMSE MAE k = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20 WNMF 1.551 ± 0.009 1.533 ± 0.009 1.156 ± 0.008 1.153 ± 0.009 1.596 ± 0.016 1.594 ± 0.016 1.219 ± 0.016 1.218 ± 0.013 Hete-MF 1.402 ± 0.010 1.398 ± 0.009 1.034 ± 0.009 1.030 ± 0.008 1.462 ± 0.017 1.454 ± 0.014 1.106 ± 0.016 1.101 ± 0.013 Hete-CF 1.148 ± 0.008 1.141 ± 0.008 0.908 ± 0.007 0.901 ± 0.007 1.211 ± 0.011 1.201 ± 0.011 0.961 ± 0.009 0.955 ± 0.009 Hete-PRW 1.395 ± 0.009 1.392 ± 0.009 1.039 ± 0.005 1.030 ± 0.005 1.434 ± 0.009 1.428 ± 0.009 1.072 ± 0.005 1.066 ± 0.005 Amp-MF 1.099 ± 0.009 1.097 ± 0.009 0.869 ± 0.005 0.868 ± 0.005 1.131 ± 0.009 1.128 ± 0.009 0.899 ± 0.005 0.897 ± 0.005 CMF 1.152 ± 0.007 1.143 ± 0.007 0.870 ± 0.004 0.868 ± 0.005 1.198 ± 0.012 1.185 ± 0.010 0.902 ± 0.009 0.899 ± 0.009 RMGM 1.246 ± 0.008 1.242 ± 0.010 0.989 ± 0.005 0.983 ± 0.007 1.271 ± 0.005 1.266 ± 0.009 1.013 ± 0.002 1.014 ± 0.006 Amp-CMF 1.097 ± 0.009 1.095 ± 0.009 0.867 ± 0.005 0.866 ± 0.005 1.129 ± 0.009 1.127 ± 0.009 0.898 ± 0.005 0.896 ± 0.005 Method RMSE MAE RMSE MAE k = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20 WNMF 1.446 ± 0.009 1.429 ± 0.009 1.097 ± 0.006 1.083 ± 0.005 1.535 ± 0.014 1.520 ± 0.013 1.184 ± 0.005 1.170 ± 0.005 Hete-MF 1.429 ± 0.009 1.351 ± 0.009 1.086 ± 0.005 1.006 ± 0.005 1.518 ± 0.012 1.492 ± 0.011 1.171 ± 0.005 1.148 ± 0.005 Hete-CF 1.305 ± 0.008 1.199 ± 0.008 0.957 ± 0.005 0.907 ± 0.005 1.378 ± 0.009 1.228 ± 0.009 1.017 ± 0.005 0.935 ± 0.005 Hete-PRW 1.343 ± 0.008 1.313 ± 0.008 1.018 ± 0.005 0.991 ± 0.005 1.414 ± 0.008 1.382 ± 0.008 1.088 ± 0.005 1.059 ± 0.005 Amp-MF 1.191 ± 0.009 1.187 ± 0.009 0.899 ± 0.005 0.897 ± 0.005 1.219 ± 0.008 1.215 ± 0.008 0.928 ± 0.005 0.925 ± 0.005 CMF 1.294 ± 0.009 1.274 ± 0.010 0.966 ± 0.005 0.949 ± 0.006 1.349 ± 0.012 1.329 ± 0.012 1.015 ± 0.005 0.998 ± 0.006 RMGM 1.240 ± 0.009 1.238 ± 0.009 0.925 ± 0.005 0.902 ± 0.004 1.316 ± 0.009 1.295 ± 0.009 0.995 ± 0.004 0.974 ± 0.004 Amp-CMF 1.134 ± 0.009 1.127 ± 0.009 0.854 ± 0.005 0.847 ± 0.005 1.148 ± 0.009 1.139 ± 0.009 0.875 ± 0.006 0.865 ± 0.006 Name #users #items #tags #ratings #social links Yelp 26, 618 8, 467 770 230, 418 183, 765 Epinions 21, 740 31, 678 27 541, 108 344, 286

Transcript of Author Title Author Title Author Item Recommendation for ...clu/doc/ijcai16_AmpCMF_poster.pdf · m...

Page 1: Author Title Author Title Author Item Recommendation for ...clu/doc/ijcai16_AmpCMF_poster.pdf · m ⇡] 2 R m⇡ ⇥k are low-rank representation of users and items. n⇡ and m⇡

Item Recommendation for Emerging Online BusinessesChun-Ta Lu ([email protected]), Sihong Xie, Weixiang Shao, Lifang He and Philip S. Yu

1.Item Recommendation for Emerging Businesses

(I) Emerging Business: new network with very sparse information.(II) Partially Aligned Networks: networks sharing common entities

crowd-sourcedreview site

? ?

followreview

aligned entity

5

4

4

5

53

?

3

4

C1 C2 C3 C4

P1 P2 P3

U1 U2

U3

U4

U5

I1 I2 I3

I4

4

?

electronic

art

sport

ratetag

E-commerce site

How to measure similarity between users/items with sparse information? How to utilize information from aligned networks to improve performance?

Hint: Transfer knowledge through aligned entities.

P

5.AcknowledgmentsThis work is supported in part by NSF through grants III-1526499, and CNS-1115234, NSFC through grant 61503253, and Google Research Award.

2. Augmented Meta Path-based Similarity (AmpSim)

Product (P)

Customer (C) User (U)

Item (I)

Rate Review

Score

Star

Follow

Partially aligned

Partially aligned

crowd-sourced review site

(network )

Tag (T)

Has

e-commerce site

(network ) G(s)G(t)

Schema of partially aligned networks

Examples of Augmented Meta Path (AMP)

Normalize link attributes to deal with different types of link attributes upon an AMP:

Average over multiple AMPs:

How to measure similarity via AMPs?

Table 5: overall Yelp dataset

MethodRMSE MAE

K = 10 K = 20 K = 10 K = 20WNMF 1.446± 0.009 1.429± 0.009 1.097± 0.006 1.083± 0.005Hete-MF 1.518± 0.012 1.492± 0.011 1.171± 0.005 1.148± 0.005Hete-CF 1.305± 0.008 1.199± 0.008 0.957± 0.005 0.907± 0.005Hete-PRW 1.343± 0.008 1.313± 0.008 1.018± 0.005 0.991± 0.005Amp-MF 1.191± 0.009 1.187± 0.009 0.899± 0.005 0.897± 0.005CMF 1.294± 0.009 1.274± 0.010 0.966± 0.005 0.949± 0.006RMGM 1.240± 0.009 1.238± 0.009 0.925± 0.005 0.902± 0.004Amp-CMF 1.134± 0.009 1.127± 0.009 0.854± 0.005 0.847± 0.005

Table 6: Cold-Start Yelp dataset

MethodRMSE MAE

K = 10 K = 20 K = 10 K = 20WNMF 1.535± 0.014 1.520± 0.013 1.184± 0.005 1.170± 0.005Hete-MF 1.462± 0.017 1.454± 0.014 1.106± 0.016 1.101± 0.013Hete-CF 1.378± 0.009 1.228± 0.009 1.017± 0.005 0.935± 0.005Hete-PRW 1.414± 0.008 1.382± 0.008 1.088± 0.005 1.059± 0.005Amp-MF 1.219± 0.008 1.215± 0.008 0.928± 0.005 0.925± 0.005CMF 1.349± 0.012 1.329± 0.012 1.015± 0.005 0.998± 0.006RMGM 1.316± 0.009 1.295± 0.009 0.995± 0.004 0.974± 0.004Amp-CMF 1.148± 0.009 1.139± 0.009 0.875± 0.006 0.865± 0.006

Table 7: Augmented meta paths used in the experiment.Intra-network augmented meta path

(C[star]���! P

[star] ��� C),(P[star] ��� C

[star]���! P ), (P ! T P )

(C[star]���! P

[star] ��� C[star]���! P

[star] ��� C)

(P[star] ��� C

[star]���! P[star] ��� C

[star]���! P )Inter-network augmented meta path

(P ! I[score] ���� U

[score]����! I ! P )

(C ! U[score]����! I

[score] ���� U ! C), (C ! U ! U ! C)

3

both linkage structures and augmented link attributes are taken into account

Product (P)

Customer (C) User (U)

Item (I)

Rate Review

Score

Star

Follow

Partially aligned

Partially aligned

crowd-sourced review site

(network )

Tag (T)

Has

e-commerce site

(network ) G(s)G(t)

Figure 2: Schema of an e-commerce site partially alignedwith a crowd-sourced review site.

For simplicity, we will use the first letter of each entityto represent that entity, the opposite direction of the arrowto represent the reverse relation, ignore the relation type ifthere is no ambiguity, and ignore the link attribute type if itis simply the count of connections. If an AMP P does notinvolve any anchor link relations RA, we call it as an intra-network AMP. Otherwise, we call it as an inter-network AMP.Example 1. Intra-network AMP: The products rated by thesame customer can be captured by P

[star] ��� C[star]���! P .

Example 2. Inter-network AMP: The products in G(t) thatare reviewed by the same users in G(s) can be observed byP ! I

[score] ��� U[score]���! I ! P , where ! denotes the

anchor links across G(s) and G(t).Inspired by the Homophily Principle [McPherson et al.,

2001] – two entities are considered to be similar if they arerelated to similar neighbors – we propose to measure the sim-ilarity of two entities by their neighbors’ similarities. How-ever, an AMP may consist of multiple types of relations, anddifferent link attributes are hard to be compared. To providea general similarity measure, we apply a simple trick to nor-malize the link attributes in the AMP.

Let first assume that in the AMP the link attributes appearin pairs (i.e., the number of the same type of link attributes iseven). We can replace the link attribute Y by the normalizedvalue M using following equation:

Mui

=Yui

� biqP

j

(Yuj

� bj

)2, (1)

where bi

is a bias term for the entity i. For the link attributesthat are bounded within a certain range (e.g., a rating can bebounded between 1 and 5), we set b

i

= Yi

the average valueof entity i. For the other type of attributes, we set b

i

= 0. It isnot hard to see that the multiplication of a pair of normalizedlink attributes (i.e., MMT ) equals to adjusted cosine similar-ity if b

i

6= 0 and equals to cosine similarity if bi

= 0. Forthe link attributes that do not appear in pairs, there is no wayto compare the link attributes with others. Thus, we replacethese attributes by the count of the connections and normalizethem using Equation 1 with b

i

= 0.Considering the normalized link attributes and the direc-

tion of AMPs, we formulate the similarity measure as:

Definition 4. AmpSim: An augmented meta path-based sim-ilarity measure between v

a

and vb

based on path P is:

s(va

, vb

|P) =[Q

l

i=1 Mi

]ab

+ [Q1

i=l

MT

i

]ba

[Q

l

i=1 Mi

]a⇤ + [

Q1i=l

MT

i

]b⇤2 [0, 1],

whereQ

l

i=1 Mi

denotes the product of the normalized linkattributes upon P , [M ]

ab

means the entry (a, b) in M, and[M ]

a⇤ means the ath row in M. Since AmpSim can becomputed in matrix form, the time complexity of computingAmpSim equals to that of matrix multiplications. Note thatwhen P is symmetric and all the link attributes are the countof connections, AmpSim is equivalent to PathSim [Sun et al.,2011]. Hence, AmpSim can be seen as a generalized versionof PathSim on AMPs.

Different AMPs capture the similarity between entities indifferent aspects and overall similarity between entities canbe obtained by aggregating information from all these AMPs.Without loss of generality, we choose tanh as the aggregationfunction, the overall similarity between entity v

a

and vb

canbe represented as

S(va

, vb

) =tanh(

Pi

wi

Si

(va

, vb

))

tanh(1)2 [0, 1], (2)

where Si

denotes the AmpSim upon the AMP Pi

, the value ofw

i

denotes the weight of Pi

, andP

i

wi

= 1.

4 Recommendation ModelIn transfer learning based collaborative filtering, collectivematrix factorization (CMF) [Singh and Gordon, 2008] is pro-posed to estimate the low-rank representations as follows:

minP(⇡)

,Q(⇡),⇡2{s,t}

X

⇡2{s,t}

||I(⇡) � (Y(⇡) �P(⇡)Q(⇡)T )||2F

,

where � is the Hadamard (element-wise) product, || · ||F

stands for Frobenius norm, and I(⇡) is an indicator ma-trix. I(⇡)

ui

= 1 if Y(⇡)ui

is observed, and otherwise I(⇡)ui

=

0. P(⇡) = [p(⇡)1 ;p(⇡)

2 ; . . . ;p(⇡)n⇡ ] 2 Rn⇡⇥k and Q(⇡) =

[q(⇡)1 ;q(⇡)

2 ; . . . ;q(⇡)m⇡ ] 2 Rm⇡⇥k are low-rank representation

of users and items. n⇡

and m⇡

are the number of users anditems in network G(⇡), and k is the parameter that estimatethe rank. The key idea of CMF is to share parameters amongfactors when an entity participates in multiple relations. In[Singh and Gordon, 2008], for instance, the factor matricesof items are assumed to be the same (i.e., Q(s) = Q(t)).

We formulate our model as a constrained collective matrixfactorization, where three soft constraints are involved:

• Non-negativity: The factor matrices {P(⇡)}, {Q(⇡)}contain only nonnegative entries, since we only focuson positive interactions between users and items.

• Geometric closeness: The latent factors of similar enti-ties should be close w.r.t. the geometric structure. Pre-serving the geometric structure in the target domain canbe achieved by the geometric regularization [Cai et al.,2011]:

RG

(P) =1

2

X

u,v

Suv

||pu

� pv

||22,

is the product of the normalized link attributes upon

Ordinal attributes: bi = Yi

Other attributes: bi = 0MMT is (adjusted) cosine similarityAfter normalization,

AmpSim: AMP-based similarity measure between va and vbbased on path

Y: link attributesMui =

Yui � biqPj(Yuj � bj)2

s(va, vb|P) =[Ql

i=1 Mi]ab + [Q1

i=l MTi ]ba

[Ql

i=1 Mi]a⇤ + [Q1

i=l MTi ]b⇤

2 [0, 1]

S(va, vb) =tanh(

Pi wiSi(va, vb))

tanh(1)2 [0, 1]

path direction is taken into account

Product (P)

Customer (C) User (U)

Item (I)

Rate Review

Score

Star

Follow

Partially aligned

Partially aligned

crowd-sourced review site

(network )

Tag (T)

Has

e-commerce site

(network ) G(s)G(t)

Figure 2: Schema of an e-commerce site partially alignedwith a crowd-sourced review site.

For simplicity, we will use the first letter of each entityto represent that entity, the opposite direction of the arrowto represent the reverse relation, ignore the relation type ifthere is no ambiguity, and ignore the link attribute type if itis simply the count of connections. If an AMP P does notinvolve any anchor link relations RA, we call it as an intra-network AMP. Otherwise, we call it as an inter-network AMP.Example 1. Intra-network AMP: The products rated by thesame customer can be captured by P

[star] ��� C[star]���! P .

Example 2. Inter-network AMP: The products in G(t) thatare reviewed by the same users in G(s) can be observed byP ! I

[score] ��� U[score]���! I ! P , where ! denotes the

anchor links across G(s) and G(t).Inspired by the Homophily Principle [McPherson et al.,

2001] – two entities are considered to be similar if they arerelated to similar neighbors – we propose to measure the sim-ilarity of two entities by their neighbors’ similarities. How-ever, an AMP may consist of multiple types of relations, anddifferent link attributes are hard to be compared. To providea general similarity measure, we apply a simple trick to nor-malize the link attributes in the AMP.

Let first assume that in the AMP the link attributes appearin pairs (i.e., the number of the same type of link attributes iseven). We can replace the link attribute Y by the normalizedvalue M using following equation:

Mui

=Yui

� biqP

j

(Yuj

� bj

)2, (1)

where bi

is a bias term for the entity i. For the link attributesthat are bounded within a certain range (e.g., a rating can bebounded between 1 and 5), we set b

i

= Yi

the average valueof entity i. For the other type of attributes, we set b

i

= 0. It isnot hard to see that the multiplication of a pair of normalizedlink attributes (i.e., MMT ) equals to adjusted cosine similar-ity if b

i

6= 0 and equals to cosine similarity if bi

= 0. Forthe link attributes that do not appear in pairs, there is no wayto compare the link attributes with others. Thus, we replacethese attributes by the count of the connections and normalizethem using Equation 1 with b

i

= 0.Considering the normalized link attributes and the direc-

tion of AMPs, we formulate the similarity measure as:

Definition 4. AmpSim: An augmented meta path-based sim-ilarity measure between v

a

and vb

based on path P is:

s(va

, vb

|P) =[Q

l

i=1 Mi

]ab

+ [Q1

i=l

MT

i

]ba

[Q

l

i=1 Mi

]a⇤ + [

Q1i=l

MT

i

]b⇤2 [0, 1],

whereQ

l

i=1 Mi

denotes the product of the normalized linkattributes upon P , [M ]

ab

means the entry (a, b) in M, and[M ]

a⇤ means the ath row in M. Since AmpSim can becomputed in matrix form, the time complexity of computingAmpSim equals to that of matrix multiplications. Note thatwhen P is symmetric and all the link attributes are the countof connections, AmpSim is equivalent to PathSim [Sun et al.,2011]. Hence, AmpSim can be seen as a generalized versionof PathSim on AMPs.

Different AMPs capture the similarity between entities indifferent aspects and overall similarity between entities canbe obtained by aggregating information from all these AMPs.Without loss of generality, we choose tanh as the aggregationfunction, the overall similarity between entity v

a

and vb

canbe represented as

S(va

, vb

) =tanh(

Pi

wi

Si

(va

, vb

))

tanh(1)2 [0, 1], (2)

where Si

denotes the AmpSim upon the AMP Pi

, the value ofw

i

denotes the weight of Pi

, andP

i

wi

= 1.

4 Recommendation ModelIn transfer learning based collaborative filtering, collectivematrix factorization (CMF) [Singh and Gordon, 2008] is pro-posed to estimate the low-rank representations as follows:

minP(⇡)

,Q(⇡),⇡2{s,t}

X

⇡2{s,t}

||I(⇡) � (Y(⇡) �P(⇡)Q(⇡)T )||2F

,

where � is the Hadamard (element-wise) product, || · ||F

stands for Frobenius norm, and I(⇡) is an indicator ma-trix. I(⇡)

ui

= 1 if Y(⇡)ui

is observed, and otherwise I(⇡)ui

=

0. P(⇡) = [p(⇡)1 ;p(⇡)

2 ; . . . ;p(⇡)n⇡ ] 2 Rn⇡⇥k and Q(⇡) =

[q(⇡)1 ;q(⇡)

2 ; . . . ;q(⇡)m⇡ ] 2 Rm⇡⇥k are low-rank representation

of users and items. n⇡

and m⇡

are the number of users anditems in network G(⇡), and k is the parameter that estimatethe rank. The key idea of CMF is to share parameters amongfactors when an entity participates in multiple relations. In[Singh and Gordon, 2008], for instance, the factor matricesof items are assumed to be the same (i.e., Q(s) = Q(t)).

We formulate our model as a constrained collective matrixfactorization, where three soft constraints are involved:

• Non-negativity: The factor matrices {P(⇡)}, {Q(⇡)}contain only nonnegative entries, since we only focuson positive interactions between users and items.

• Geometric closeness: The latent factors of similar enti-ties should be close w.r.t. the geometric structure. Pre-serving the geometric structure in the target domain canbe achieved by the geometric regularization [Cai et al.,2011]:

RG

(P) =1

2

X

u,v

Suv

||pu

� pv

||22,

va and vband

P

3. Recommendation Model: AmpCMF

Collective Matrix Factorization (CMF) w/ alignment regularization

geometric regularization

Title

Author

1 MFM in matrix form

Let Z(p) 2 RN⇥(1+dp) be the feature matrix of the p-th view, IT 2 RN⇥T be thetask indictor matrix, and I = [1 IT ] 2 RN⇥(1+T ). Then the predictive valuesof all the instances can be computed in matrix form:

y =⇣⇣

Z(1)⇥(1)⌘� · · ·�

⇣Z(V )⇥(V )

⌘⌘(I�)

T , (1)

where yi is the i-th instance.

minP(⇡),Q(⇡)�0,⇡2{s,t}

J =X

⇡2{s,t}

||I(⇡) � (Y(⇡) �P(⇡)Q(⇡)T )||2F

+ �(RA(P(t),P(s)) +RA(Q

(t),Q(s)))

+ �(RG(P(t)) +RG(Q

(t)))

+↵

2

X

⇡2{s,t}

(||P(⇡)||2F + ||Q(⇡)||2F )

1

Title

Author

1 MFM in matrix form

Let Z(p) 2 RN⇥(1+dp) be the feature matrix of the p-th view, IT 2 RN⇥T be thetask indictor matrix, and I = [1 IT ] 2 RN⇥(1+T ). Then the predictive valuesof all the instances can be computed in matrix form:

y =⇣⇣

Z(1)⇥(1)⌘� · · ·�

⇣Z(V )⇥(V )

⌘⌘(I�)

T , (1)

where yi is the i-th instance.

minP(⇡),Q(⇡)�0,⇡2{s,t}

J =X

⇡2{s,t}

||I(⇡) � (Y(⇡) �P(⇡)Q(⇡)T )||2F

+ �(RA(P(t),P(s)) +RA(Q

(t),Q(s)))

+ �(RG(P(t)) +RG(Q

(t)))

+↵

2

X

⇡2{s,t}

(||P(⇡)||2F + ||Q(⇡)||2F )

1

Title

Author

1 MFM in matrix form

Let Z(p) 2 RN⇥(1+dp) be the feature matrix of the p-th view, IT 2 RN⇥T be thetask indictor matrix, and I = [1 IT ] 2 RN⇥(1+T ). Then the predictive valuesof all the instances can be computed in matrix form:

y =⇣⇣

Z(1)⇥(1)⌘� · · ·�

⇣Z(V )⇥(V )

⌘⌘(I�)

T , (1)

where yi is the i-th instance.

minP(⇡),Q(⇡)�0,⇡2{s,t}

J =X

⇡2{s,t}

||I(⇡) � (Y(⇡) �P(⇡)Q(⇡)T )||2F

+ �(RA(P(t),P(s)) +RA(Q

(t),Q(s)))

+ �(RG(P(t)) +RG(Q

(t)))

+↵

2

X

⇡2{s,t}

(||P(⇡)||2F + ||Q(⇡)||2F )

1

prevent overfitting

alignment regularization:

geometric regularization:

✦ Solve the optimization problem using multiplicative updates

Y and I 2 Rn⇥m are user-item feedback and indicator matricesP 2 Rn⇥k and Q 2 Rm⇥k are low rank representations of users and items

RG(P) =1

2

X

u,v

Suv||pu � pv||22 =1

2Tr(PT (D� S)P) =

1

2Tr(PTLPP)

RA(P(t),P(s)) =

1

2

X

(i,gi)2A(t,s)

||p(t)i � p(s)

gi ||22 =

1

2kAATP(t) �AP(s)k2F

Figure 7: Illustration of Splits

Disjoint Split Overlap Split Contained Split Subset Split

training set of R1training set of R2

Partial Split

users

items# of Ratings:

40% of R50% of R20% of R

# of Ratings:40% of R50% of R0% of R

# of Users/Items:40% of R

100% of R40% of R

# of Ratings:40% of R50% of R40% of R

# of Users/Items:40% of R90% of R30% of R

Even when R1 = R2, we can still obtain very different P1 andP2. In fact, experiment results show that the first two baselinesperform no better than random guess, so we omit them from theresult table to save space. We found the third baseline that worksmuch better than the first two, and therefore its results are includedin our table.

Second, we conduct an experiment to evaluate the quality boostof rating prediction in R1 after transferring information from R2.There has not yet been much work on information transfer withoutknowing the underlying correspondence. Hence, for rating predic-tion we use rating-matrix generative model (RMGM) as the state-of-the-art for comparison.

4.3 Implementation DetailsFor our factorization models, we use gradient descent and back-

tracking line search to solve the objective function. The dimensionK is fixed to 50, and the parameters λ and β are automaticallyselected by observing the error rate of the validation set. After se-lection, λ is determined to be 0 for noise-free low-rank dataset and5 for Yahoo! music dataset, and β ranges from 0 to 400. Besides,we employ the data scaling and early stopping procedure. We scalethe ratings in the training set to zero mean and unit variance, andscale the values back in the prediction phase; the training processstops when validation error starts to increase.

We implement two versions of rating-matrix generative model(RMGM) [8] for comparison. They can be solved by standardexpectation-maximization algorithm. 2 The original RMGM usescategorical distribution for P (r|Cm, Cn). However, categoricaldistribution does not reflect the ordinal relation among the ratingvalues. Therefore, we implement another version of RMGM thatuses Gaussian distribution. For both models, the original and Gaus-sian RMGMs, we set the latent dimension K to 50 (we find that alarger K leads to similar performance), and we conduct the sameearly stopping procedure. For Gaussian RMGM, we find that thevariance of Gaussian is better set to a constant parameter that canbe tuned by observing the error rate of the validation set. We havealso conducted the data scaling process for Gaussian RMGM.

4.4 Results of User and Item MatchingIn the matching process, the matching can output either the most

likely candidate or a ranked list of candidates for each user/itemin R1. Therefore, we can use accuracy as well as mean averageprecision (MAP) as the evaluation criteria. However, in the partialsplit, some users and items in R1 do not appear in R2. Thus, when

2An example code is provided by the author of RMGM: https://sites.google.com/site/libin82cn/

Table 2: Matching Result on the Low-Rank Dataset

Rating LatentRefinement

List Space

Disjoint SplitAccuracy(user) 0.000 1.000 1.000

MAP(user) 0.003 1.000 1.000Accuracy(item) 0.001 1.000 1.000

MAP(item) 0.007 1.000 1.000Overlap SplitAccuracy(user) 0.025 1.000 1.000

MAP(user) 0.048 1.000 1.000Accuracy(item) 0.051 1.000 1.000

MAP(item) 0.089 1.000 1.000Contained SplitAccuracy(user) 0.538 1.000 1.000

MAP(user) 0.612 1.000 1.000Accuracy(item) 0.703 1.000 1.000

MAP(item) 0.765 1.000 1.000Subset Split

Accuracy(user) 0.013 1.000 1.000MAP(user) 0.029 1.000 1.000

Accuracy(item) 0.029 1.000 1.000MAP(item) 0.058 1.000 1.000

Partial SplitAccuracy(user) 0.007 1.000 1.000

MAP(user) 0.019 1.000 1.000Accuracy(item) 0.018 1.000 1.000

MAP(item) 0.042 1.000 1.000

evaluating the matching result of such scenario, we remove theseusers and items from consideration.

The matching results are shown in Table 2 and Table 3. Theresults on the noise-free low-rank dataset demonstrate the sound-ness of our approach. When there is no noise in the rating ma-trix, our latent space matching can precisely match all users anditems by comparing the singular vectors regardless of the splits.On the other hand, the performance of baseline model (rating list)improves when more ratings are shared between domains , but theresults are still far from perfect.

On Yahoo! music dataset, we see a similar trend: when more rat-ings are shared, the results are generally better. We observe that thecontained split is easier to match than overlap split, while both areeasier to match than disjoint split. The matching accuracy for thesubset split and partial split are slightly worse (though still com-parable) than that of the overlap split. It is because although the

808

4. Experiment Results

cold-start: test on users who have ≤ 5 ratings in training set90% (55%) of users are cold-start users in Yelp (Epinions)

Yelp: test a mature business entering a new domain

Figure 7: Illustration of Splits

Disjoint Split Overlap Split Contained Split Subset Split

training set of R1training set of R2

Partial Split

users

items# of Ratings:

40% of R50% of R20% of R

# of Ratings:40% of R50% of R0% of R

# of Users/Items:40% of R

100% of R40% of R

# of Ratings:40% of R50% of R40% of R

# of Users/Items:40% of R90% of R30% of R

Even when R1 = R2, we can still obtain very different P1 andP2. In fact, experiment results show that the first two baselinesperform no better than random guess, so we omit them from theresult table to save space. We found the third baseline that worksmuch better than the first two, and therefore its results are includedin our table.

Second, we conduct an experiment to evaluate the quality boostof rating prediction in R1 after transferring information from R2.There has not yet been much work on information transfer withoutknowing the underlying correspondence. Hence, for rating predic-tion we use rating-matrix generative model (RMGM) as the state-of-the-art for comparison.

4.3 Implementation DetailsFor our factorization models, we use gradient descent and back-

tracking line search to solve the objective function. The dimensionK is fixed to 50, and the parameters λ and β are automaticallyselected by observing the error rate of the validation set. After se-lection, λ is determined to be 0 for noise-free low-rank dataset and5 for Yahoo! music dataset, and β ranges from 0 to 400. Besides,we employ the data scaling and early stopping procedure. We scalethe ratings in the training set to zero mean and unit variance, andscale the values back in the prediction phase; the training processstops when validation error starts to increase.

We implement two versions of rating-matrix generative model(RMGM) [8] for comparison. They can be solved by standardexpectation-maximization algorithm. 2 The original RMGM usescategorical distribution for P (r|Cm, Cn). However, categoricaldistribution does not reflect the ordinal relation among the ratingvalues. Therefore, we implement another version of RMGM thatuses Gaussian distribution. For both models, the original and Gaus-sian RMGMs, we set the latent dimension K to 50 (we find that alarger K leads to similar performance), and we conduct the sameearly stopping procedure. For Gaussian RMGM, we find that thevariance of Gaussian is better set to a constant parameter that canbe tuned by observing the error rate of the validation set. We havealso conducted the data scaling process for Gaussian RMGM.

4.4 Results of User and Item MatchingIn the matching process, the matching can output either the most

likely candidate or a ranked list of candidates for each user/itemin R1. Therefore, we can use accuracy as well as mean averageprecision (MAP) as the evaluation criteria. However, in the partialsplit, some users and items in R1 do not appear in R2. Thus, when

2An example code is provided by the author of RMGM: https://sites.google.com/site/libin82cn/

Table 2: Matching Result on the Low-Rank Dataset

Rating LatentRefinement

List Space

Disjoint SplitAccuracy(user) 0.000 1.000 1.000

MAP(user) 0.003 1.000 1.000Accuracy(item) 0.001 1.000 1.000

MAP(item) 0.007 1.000 1.000Overlap SplitAccuracy(user) 0.025 1.000 1.000

MAP(user) 0.048 1.000 1.000Accuracy(item) 0.051 1.000 1.000

MAP(item) 0.089 1.000 1.000Contained SplitAccuracy(user) 0.538 1.000 1.000

MAP(user) 0.612 1.000 1.000Accuracy(item) 0.703 1.000 1.000

MAP(item) 0.765 1.000 1.000Subset Split

Accuracy(user) 0.013 1.000 1.000MAP(user) 0.029 1.000 1.000

Accuracy(item) 0.029 1.000 1.000MAP(item) 0.058 1.000 1.000

Partial SplitAccuracy(user) 0.007 1.000 1.000

MAP(user) 0.019 1.000 1.000Accuracy(item) 0.018 1.000 1.000

MAP(item) 0.042 1.000 1.000

evaluating the matching result of such scenario, we remove theseusers and items from consideration.

The matching results are shown in Table 2 and Table 3. Theresults on the noise-free low-rank dataset demonstrate the sound-ness of our approach. When there is no noise in the rating ma-trix, our latent space matching can precisely match all users anditems by comparing the singular vectors regardless of the splits.On the other hand, the performance of baseline model (rating list)improves when more ratings are shared between domains , but theresults are still far from perfect.

On Yahoo! music dataset, we see a similar trend: when more rat-ings are shared, the results are generally better. We observe that thecontained split is easier to match than overlap split, while both areeasier to match than disjoint split. The matching accuracy for thesubset split and partial split are slightly worse (though still com-parable) than that of the overlap split. It is because although the

808

target network (20% of ratings)

source network (80% of ratings)overlapping (20% of users; all items)

target network: Nevada w/ 11.5K users; 2.1K items source network: Arizona w/ 19.5K users; 6.3K itemsoverlapping: 4.3K users; no items

WNMF: weighted nonnegative matrix factorization Hete-MF: WNMF+ PathSim between items onlyHete-CF: WNMF + PathSim between users/itemsHete-PRW: WNMF + Pairwise RW similarity between users/itemsAmp-MF: WNMF + AmpSim between users/items

CMF: collective matrix factorization w/ alignment regularizationRMGM: rating-matrix generative modelAmp-CMF: proposed method

WNMF + geometric regularization

Transfer learning based methods

Epinions (Overall) Epinions (Cold Start)

Yelp (Overall) Yelp (Cold Start)

Epinions: test an emerging business partially aligned with a developed business

Conclusion: 1. Methods using more accurate similarity measure achieve better performance 2. Combining both geometric and alignment regularization boosts the performance 3. Learning from highly overlapped networks can achieve better improvement:

improve 21% in Yelp and 29% in Epinions

Table 4: Performance comparison (mean±std) on the Epinions dataset

MethodOverall Cold Start

RMSE MAE RMSE MAEk = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20

WNMF 1.551± 0.009 1.533± 0.009 1.156± 0.008 1.153± 0.009 1.596± 0.016 1.594± 0.016 1.219± 0.016 1.218± 0.013Hete-MF 1.402± 0.010 1.398± 0.009 1.034± 0.009 1.030± 0.008 1.462± 0.017 1.454± 0.014 1.106± 0.016 1.101± 0.013Hete-CF 1.148± 0.008 1.141± 0.008 0.908± 0.007 0.901± 0.007 1.211± 0.011 1.201± 0.011 0.961± 0.009 0.955± 0.009

Hete-PRW 1.395± 0.009 1.392± 0.009 1.039± 0.005 1.030± 0.005 1.434± 0.009 1.428± 0.009 1.072± 0.005 1.066± 0.005Amp-MF 1.099± 0.009 1.097± 0.009 0.869± 0.005 0.868± 0.005 1.131± 0.009 1.128± 0.009 0.899± 0.005 0.897± 0.005

CMF 1.152± 0.007 1.143± 0.007 0.870± 0.004 0.868± 0.005 1.198± 0.012 1.185± 0.010 0.902± 0.009 0.899± 0.009RMGM 1.246± 0.008 1.242± 0.010 0.989± 0.005 0.983± 0.007 1.271± 0.005 1.266± 0.009 1.013± 0.002 1.014± 0.006

Amp-CMF 1.097± 0.009 1.095± 0.009 0.867± 0.005 0.866± 0.005 1.129± 0.009 1.127± 0.009 0.898± 0.005 0.896± 0.005

Table 5: Performance comparison (mean±std) on the Yelp dataset

MethodOverall Cold Start

RMSE MAE RMSE MAEk = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20

WNMF 1.446± 0.009 1.429± 0.009 1.097± 0.006 1.083± 0.005 1.535± 0.014 1.520± 0.013 1.184± 0.005 1.170± 0.005Hete-MF 1.429± 0.009 1.351± 0.009 1.086± 0.005 1.006± 0.005 1.518± 0.012 1.492± 0.011 1.171± 0.005 1.148± 0.005Hete-CF 1.305± 0.008 1.199± 0.008 0.957± 0.005 0.907± 0.005 1.378± 0.009 1.228± 0.009 1.017± 0.005 0.935± 0.005

Hete-PRW 1.343± 0.008 1.313± 0.008 1.018± 0.005 0.991± 0.005 1.414± 0.008 1.382± 0.008 1.088± 0.005 1.059± 0.005Amp-MF 1.191± 0.009 1.187± 0.009 0.899± 0.005 0.897± 0.005 1.219± 0.008 1.215± 0.008 0.928± 0.005 0.925± 0.005

CMF 1.294± 0.009 1.274± 0.010 0.966± 0.005 0.949± 0.006 1.349± 0.012 1.329± 0.012 1.015± 0.005 0.998± 0.006RMGM 1.240± 0.009 1.238± 0.009 0.925± 0.005 0.902± 0.004 1.316± 0.009 1.295± 0.009 0.995± 0.004 0.974± 0.004

Amp-CMF 1.134± 0.009 1.127± 0.009 0.854± 0.005 0.847± 0.005 1.148± 0.009 1.139± 0.009 0.875± 0.006 0.865± 0.006

the effectiveness of Amp-CMF in addressing item recommen-dation for emerging businesses.

6 Related WorkTransfer learning based on CMF has been proposed to uti-lized information shared by related domains to learn latentfactors for better recommendations. The first category ofsuch methods assume that there are certain overlapping usersand/or items across multiple domains and align the latent fac-tors of the overlapping users and/or items [Singh and Gor-don, 2008; Long et al., 2010]. Another approach is to en-force the similarity of cluster-level preference patterns in tworelated domains, without assuming cross-domain entity cor-respondences. Codebook-based-transfer (CBF) [Li et al.,2009a] proposed to align the two kernel matrices of usersfrom two domains using a co-clustering process. Rating-matrix generative model (RMGM) relaxes the constraints inCBT from hard clustering to soft clustering by using a prob-abilistic graphical model. Although the former approachis more effective for knowledge transfer in general, it canbe biased to the overlapping portions. One possible wayto handle the above issue is to preserve geometric close-ness between similar entities [Wang and Mahadevan, 2011;Long et al., 2014]. However, it may not be trivial to find reli-able similarity information in the emerging domain.

When building a recommendation system, it is crucial tochoose a good way to measure the entity similarity. A numberof studies leverage linkage structure of a network for measur-ing similarity, e.g., personalized PageRank [Jeh and Widom,2003] and SimRank [Jeh and Widom, 2002], while they aredefined on a homogeneous network or a bipartite network.However, in real-world applications, there are lots of hetero-geneous networks. Meta path was introduced in [Sun et al.,2011; Shi et al., 2012] to measure the similarity in heteroge-neous networks. A series of meta path-based similarity mea-

sures are proposed for either the same type of entities [Sunet al., 2011] or different type of entities [Shi et al., 2014;Cao et al., 2014; Shi et al., 2015]. Recently, researchers havebeen aware of the importance of heterogeneous informationfor recommendation. Yu et al. [Yu et al., 2013] proposed animplicit feedback recommendation model with the similarityinformation extracted from heterogeneous network. Luo etal. [Luo et al., 2014] proposed a collaborative filtering-basedsocial recommendation method, called Hete-CF, using het-erogeneous relations. Vahedian [Vahedian, 2014] proposedthe WHyLDR approach for multiple recommendation tasks,which combines heterogeneous information with a linear-weighted hybrid model. These methods usually treat the re-lationships between entities as binary connections. However,plentiful attributes are usually attached to the relationships,e.g., the rating of an item given by a user and the timestamp ofa post. As demonstrated in the experiments, discarding theseimportant attributes can lead to a degraded performance.

7 ConclusionIn this paper, we have proposed a novel similarity measurecalled AmpSim that can judiciously capture the rich simi-larity semantics between entities by taking both the linkagestructures and the augmented link attributes into account.We further incorporated the similarity information capturedby AmpSim in a constrained collective matrix factorizationmodel. Extensively experiments on real-world datasets havedemonstrated that our proposed model significantly outper-forms other state-of-the-art collaborative filtering algorithmsin addressing item recommendation for emerging businesses.

AcknowledgmentsThis work is supported in part by NSF through grants III-1526499, and CNS-1115234, NSFC through grant 61503253,and Google Research Award .

Table 4: Performance comparison (mean±std) on the Epinions dataset

MethodOverall Cold Start

RMSE MAE RMSE MAEk = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20

WNMF 1.551± 0.009 1.533± 0.009 1.156± 0.008 1.153± 0.009 1.596± 0.016 1.594± 0.016 1.219± 0.016 1.218± 0.013Hete-MF 1.402± 0.010 1.398± 0.009 1.034± 0.009 1.030± 0.008 1.462± 0.017 1.454± 0.014 1.106± 0.016 1.101± 0.013Hete-CF 1.148± 0.008 1.141± 0.008 0.908± 0.007 0.901± 0.007 1.211± 0.011 1.201± 0.011 0.961± 0.009 0.955± 0.009

Hete-PRW 1.395± 0.009 1.392± 0.009 1.039± 0.005 1.030± 0.005 1.434± 0.009 1.428± 0.009 1.072± 0.005 1.066± 0.005Amp-MF 1.099± 0.009 1.097± 0.009 0.869± 0.005 0.868± 0.005 1.131± 0.009 1.128± 0.009 0.899± 0.005 0.897± 0.005

CMF 1.152± 0.007 1.143± 0.007 0.870± 0.004 0.868± 0.005 1.198± 0.012 1.185± 0.010 0.902± 0.009 0.899± 0.009RMGM 1.246± 0.008 1.242± 0.010 0.989± 0.005 0.983± 0.007 1.271± 0.005 1.266± 0.009 1.013± 0.002 1.014± 0.006

Amp-CMF 1.097± 0.009 1.095± 0.009 0.867± 0.005 0.866± 0.005 1.129± 0.009 1.127± 0.009 0.898± 0.005 0.896± 0.005

Table 5: Performance comparison (mean±std) on the Yelp dataset

MethodOverall Cold Start

RMSE MAE RMSE MAEk = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20

WNMF 1.446± 0.009 1.429± 0.009 1.097± 0.006 1.083± 0.005 1.535± 0.014 1.520± 0.013 1.184± 0.005 1.170± 0.005Hete-MF 1.429± 0.009 1.351± 0.009 1.086± 0.005 1.006± 0.005 1.518± 0.012 1.492± 0.011 1.171± 0.005 1.148± 0.005Hete-CF 1.305± 0.008 1.199± 0.008 0.957± 0.005 0.907± 0.005 1.378± 0.009 1.228± 0.009 1.017± 0.005 0.935± 0.005

Hete-PRW 1.343± 0.008 1.313± 0.008 1.018± 0.005 0.991± 0.005 1.414± 0.008 1.382± 0.008 1.088± 0.005 1.059± 0.005Amp-MF 1.191± 0.009 1.187± 0.009 0.899± 0.005 0.897± 0.005 1.219± 0.008 1.215± 0.008 0.928± 0.005 0.925± 0.005

CMF 1.294± 0.009 1.274± 0.010 0.966± 0.005 0.949± 0.006 1.349± 0.012 1.329± 0.012 1.015± 0.005 0.998± 0.006RMGM 1.240± 0.009 1.238± 0.009 0.925± 0.005 0.902± 0.004 1.316± 0.009 1.295± 0.009 0.995± 0.004 0.974± 0.004

Amp-CMF 1.134± 0.009 1.127± 0.009 0.854± 0.005 0.847± 0.005 1.148± 0.009 1.139± 0.009 0.875± 0.006 0.865± 0.006

the effectiveness of Amp-CMF in addressing item recommen-dation for emerging businesses.

6 Related WorkTransfer learning based on CMF has been proposed to uti-lized information shared by related domains to learn latentfactors for better recommendations. The first category ofsuch methods assume that there are certain overlapping usersand/or items across multiple domains and align the latent fac-tors of the overlapping users and/or items [Singh and Gor-don, 2008; Long et al., 2010]. Another approach is to en-force the similarity of cluster-level preference patterns in tworelated domains, without assuming cross-domain entity cor-respondences. Codebook-based-transfer (CBF) [Li et al.,2009a] proposed to align the two kernel matrices of usersfrom two domains using a co-clustering process. Rating-matrix generative model (RMGM) relaxes the constraints inCBT from hard clustering to soft clustering by using a prob-abilistic graphical model. Although the former approachis more effective for knowledge transfer in general, it canbe biased to the overlapping portions. One possible wayto handle the above issue is to preserve geometric close-ness between similar entities [Wang and Mahadevan, 2011;Long et al., 2014]. However, it may not be trivial to find reli-able similarity information in the emerging domain.

When building a recommendation system, it is crucial tochoose a good way to measure the entity similarity. A numberof studies leverage linkage structure of a network for measur-ing similarity, e.g., personalized PageRank [Jeh and Widom,2003] and SimRank [Jeh and Widom, 2002], while they aredefined on a homogeneous network or a bipartite network.However, in real-world applications, there are lots of hetero-geneous networks. Meta path was introduced in [Sun et al.,2011; Shi et al., 2012] to measure the similarity in heteroge-neous networks. A series of meta path-based similarity mea-

sures are proposed for either the same type of entities [Sunet al., 2011] or different type of entities [Shi et al., 2014;Cao et al., 2014; Shi et al., 2015]. Recently, researchers havebeen aware of the importance of heterogeneous informationfor recommendation. Yu et al. [Yu et al., 2013] proposed animplicit feedback recommendation model with the similarityinformation extracted from heterogeneous network. Luo etal. [Luo et al., 2014] proposed a collaborative filtering-basedsocial recommendation method, called Hete-CF, using het-erogeneous relations. Vahedian [Vahedian, 2014] proposedthe WHyLDR approach for multiple recommendation tasks,which combines heterogeneous information with a linear-weighted hybrid model. These methods usually treat the re-lationships between entities as binary connections. However,plentiful attributes are usually attached to the relationships,e.g., the rating of an item given by a user and the timestamp ofa post. As demonstrated in the experiments, discarding theseimportant attributes can lead to a degraded performance.

7 ConclusionIn this paper, we have proposed a novel similarity measurecalled AmpSim that can judiciously capture the rich simi-larity semantics between entities by taking both the linkagestructures and the augmented link attributes into account.We further incorporated the similarity information capturedby AmpSim in a constrained collective matrix factorizationmodel. Extensively experiments on real-world datasets havedemonstrated that our proposed model significantly outper-forms other state-of-the-art collaborative filtering algorithmsin addressing item recommendation for emerging businesses.

AcknowledgmentsThis work is supported in part by NSF through grants III-1526499, and CNS-1115234, NSFC through grant 61503253,and Google Research Award .

Table 1: Statistics of DatasetsName #users #items #tags #ratings #social linksYelp 26, 618 8, 467 770 230, 418 183, 765Epinions 21, 740 31, 678 27 541, 108 344, 286

Table 2: overall Epinion dataset

MethodRMSE MAE

k = 10 k = 20 k = 10 k = 20WNMF 1.551± 0.009 1.533± 0.009 1.156± 0.008 1.153± 0.009Hete-MF 1.402± 0.010 1.398± 0.009 1.034± 0.009 1.030± 0.008Hete-CF 1.148± 0.008 1.141± 0.008 0.908± 0.007 0.901± 0.007Hete-PRW 1.395± 0.009 1.392± 0.009 1.039± 0.005 1.030± 0.005Amp-MF 1.099± 0.009 1.097± 0.009 0.869± 0.005 0.868± 0.005CMF 1.152± 0.007 1.143± 0.007 0.870± 0.004 0.868± 0.005RMGM 1.246± 0.008 1.242± 0.010 0.989± 0.005 0.983± 0.007Amp-CMF 1.097± 0.009 1.095± 0.009 0.867± 0.005 0.866± 0.005

Table 3: Cold-Start Epinion dataset

MethodRMSE MAE

k = 10 k = 20 k = 10 k = 20WNMF 1.596± 0.016 1.594± 0.016 1.219± 0.016 1.218± 0.013Hete-MF 1.462± 0.017 1.454± 0.014 1.106± 0.016 1.101± 0.013Hete-CF 1.211± 0.011 1.201± 0.011 0.961± 0.009 0.955± 0.009Hete-PRW 1.434± 0.009 1.428± 0.009 1.072± 0.005 1.066± 0.005Amp-MF 1.131± 0.009 1.128± 0.009 0.899± 0.005 0.897± 0.005CMF 1.198± 0.012 1.185± 0.010 0.902± 0.009 0.899± 0.009RMGM 1.271± 0.005 1.266± 0.009 1.013± 0.002 1.014± 0.006Amp-CMF 1.129± 0.009 1.127± 0.009 0.898± 0.005 0.896± 0.005

Table 4: Performance comparison (mean±std) on the Yelp dataset

Method

Overall Cold StartRMSE MAE RMSE MAE

k = 10 k = 20 k = 10 k = 20 k = 10 k = 20 k = 10 k = 20WNMF 1.446± 0.009 1.429± 0.009 1.097± 0.006 1.083± 0.005 1.535± 0.014 1.520± 0.013 1.184± 0.005 1.170± 0.005Hete-MF 1.429± 0.009 1.351± 0.009 1.086± 0.005 1.006± 0.005 1.518± 0.012 1.492± 0.011 1.171± 0.005 1.148± 0.005Hete-CF 1.305± 0.008 1.199± 0.008 0.957± 0.005 0.907± 0.005 1.378± 0.009 1.228± 0.009 1.017± 0.005 0.935± 0.005Hete-PRW 1.343± 0.008 1.313± 0.008 1.018± 0.005 0.991± 0.005 1.414± 0.008 1.382± 0.008 1.088± 0.005 1.059± 0.005Amp-MF 1.191± 0.009 1.187± 0.009 0.899± 0.005 0.897± 0.005 1.219± 0.008 1.215± 0.008 0.928± 0.005 0.925± 0.005CMF 1.294± 0.009 1.274± 0.010 0.966± 0.005 0.949± 0.006 1.349± 0.012 1.329± 0.012 1.015± 0.005 0.998± 0.006

RMGM 1.240± 0.009 1.238± 0.009 0.925± 0.005 0.902± 0.004 1.316± 0.009 1.295± 0.009 0.995± 0.004 0.974± 0.004Amp-CMF 1.134± 0.009 1.127± 0.009 0.854± 0.005 0.847± 0.005 1.148± 0.009 1.139± 0.009 0.875± 0.006 0.865± 0.006

2