The paper studies a behavioral learning process where an agent plays, at each period, an action with a probability which is proportional to the cumulative utility he got in the past with that action. The so-called CPR learning rule and the dynamic process it induces are formally stated and compared to other reinforcement rules as well as to fictitious play or the replicator dynamics.