Data-centric affect modeling may render itself restrictive in practical applications for three reasons, namely, it falls short of feature optimization, infers discrete affect classes, and deals with relatively small to average sized datasets. Though it seems practical to use the feature combinations already associated to commonly investigated sensors, there may be other potentially optimal features that can lead to new relations. Secondly, although it seems more realistic to view affect as continuous, it requires using continuous labels that will increase the difficulty of modeling. Lastly, although a large scale dataset reflects a more precise range of values for any given feature, it severely hinders computational efficiency. We address these problems when inferring physiology-affect relations from datasets that contain 2-3 million feature vectors, each with 49 features and labelled with continuous affect values. We employ automatic feature selection to acquire near optimal feature subsets and a fast approximate kNN algorithm to solve the regression problem and cope with the challenge of a large scale dataset. Our results show that high estimation accuracy may be achieved even when the selected feature subset is only about 7% of the original features. May the results here motivate the HCI community to pursue affect modeling without being deterred by large datasets and further the discussions on acquiring optimal features for accurate continuous affect approximation.