%% This BibTeX bibliography file was created using BibDesk.
%% http://bibdesk.sourceforge.net/


%% Created for Jacky Keung at 2010-08-20 15:41:26 +1000 


%% Saved with string encoding Unicode (UTF-8) 


@article{Auer2006,
	Author = {Auer, Martin and Trendowicz, Adam and Graser, Bernhard and Haunschmid, Ernst and Biffl, Stefan},
	Issue = {2},
	Journal = {IEEE Transactions on Software Engineering},
	Mendeley-Groups = {Automatically Imported,Cost},
	Pages = {83--92},
	Title = {Optimal Project Feature Weights in Analogy-Based Cost Estimation: Improvement and Limitations},
	Volume = {32},
	Year = {2006}}
@article{Kirsopp2003,
	Author = {Kirsopp, C. and Shepperd, M. and Premrag, R.},
	File = {:C$\backslash$:/Users/ekrem/Documents/My Projects/Tez Related/2009-11-Papers/ES2002\_final.pdf:pdf},
	Journal = {Research and development in intelligent systems XIX: proceedings of ES2002, the twenty-second SGAI International Conference on Knowledge Based Systems and Applied Artificial Intelligence},
	Mendeley-Groups = {Automatically Imported},
	Owner = {ekrem},
	Pages = {61},
	Publisher = {Springer-Verlag New York Inc},
	Timestamp = {2009.11.26},
	Title = {Case and feature subset selection in case-based software project effort prediction},
	Year = {2003}}


@inproceedings{Lum2008,
	Author = {Lum, Karen and Menzies, Tim and Baker, Dan},
	Booktitle = {International Society of Parametric Analysis Conference (ISPA / SCEA)},
	Date-Added = {2010-08-20 15:38:45 +1000},
	Date-Modified = {2010-08-20 15:40:59 +1000},
	Month = {May},
	Title = {2CEE, A Twenty First Century Effort Estimation Methodology},
	Year = {2008}}

@article{menzies11,
	Abstract = {There exists a large and growing number of proposed estimation methods but little conclusive evidence ranking one method over another. Prior effort estimation studies suffered from  conclusion instability , where the rankings offered to different methods were not stable across (a) different evaluation criteria; (b) different data sources; or (c) different random selections of that data. This paper reports a study of 158 effort estimation methods on data sets based on COCOMO features. Four  best  methods were detected that were consistently better than the  rest  of the other 154 methods. These rankings of  best  and  rest  methods were stable across (a) three different evaluation criteria applied to (b) multiple data sets from two different sources that were (c) divided into hundreds of randomly selected subsets using four different random seeds. Hence, while there exists no single universal  best  effort estimation method, there appears to exist a small number (four) of most useful methods. This result both complicates and simplifies effort estimation research. The complication is that any future effort estimation analysis should be preceded by a  selection study  that finds the best local estimator. However, the simplification is that such a study need not be labor intensive, at least for COCOMO style data sets.},
	Affiliation = {West Virginia University Lane Department of Computer Science and Electrical Engineering Morgantown USA},
	Author = {Menzies, Tim and Jalali, Omid and Hihn, Jairus and Baker, Dan and Lum, Karen},
	Date-Added = {2010-08-20 15:34:05 +1000},
	Date-Modified = {2010-08-20 15:36:00 +1000},
	Issn = {0928-8910},
	Issue = {4},
	Journal = {Automated Software Engineering},
	Keyword = {Computer Science},
	Pages = {409-437},
	Publisher = {Springer Netherlands},
	Title = {Stable rankings for different effort models},
	Url = {http://dx.doi.org/10.1007/s10515-010-0070-z},
	Volume = {17},
	Year = {2010},
	Bdsk-Url-1 = {http://dx.doi.org/10.1007/s10515-010-0070-z}}

@inproceedings{MenziesBrady10,
	Author = {Adam Brady and Tim Menzies},
	Booktitle = {International Conference on Predictive Models in Software Engineering PROMISE'10},
	Date-Added = {2010-08-19 22:30:35 +1000},
	Date-Modified = {2010-08-19 22:35:27 +1000},
	Month = {Sept.},
	Publisher = {IEEE},
	Title = {Case-Based Reasoning vs Parametric Models for Software Quality Optimization},
	Year = {2010}}

@inproceedings{LiandRuheAtPRomise2007,
	Abstract = {Effort estimation by analogy (EBA) is an established method for software effort estimation. For this paper, we understand EBA as a meta-method which needs to be instantiated and customized at different stages and decision points regarding a specific context. Some example decision problems are related to the selection of the similarity measures, the selection of analogs for adaptation or the weighting and selection of attributes. This paper proposes a decision-centric process model for EBA by generalizing the existing EBA methods. Typical decision-making problems are identified at different stages of the process as part of the model. Some existing solution alternatives of the decision-making problems are then studied. The results of the decision support analysis can be used for better understanding of EBA related techniques and for providing guidelines for implementation and customization of general EBA. An example case of the process model is finally presented.},
	Author = {Jingzhou Li and Ruhe, G.},
	Booktitle = {International Conference on Predictive Models in Software Engineering PROMISE'07},
	Date-Modified = {2010-08-19 22:34:05 +1000},
	Doi = {10.1109/PROMISE.2007.5},
	Keywords = {decision support analysis;decision-centric process;decision-making problems;software effort estimation by analogy;decision making;decision support systems;software engineering;},
	Month = {May},
	Title = {Decision Support Analysis for Software Effort Estimation by Analogy},
	Year = {2007},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/PROMISE.2007.5}}

@article{Jor2004e,
	Author = {Magne J{\o}rgensen},
	Bibsource = {DBLP, http://dblp.uni-trier.de},
	Date-Modified = {2010-08-17 15:48:01 +1000},
	Ee = {http://dx.doi.org/10.1016/S0164-1212(02)00156-5},
	Journal = {Journal of Systems and Software},
	Number = {1-2},
	Pages = {37-60},
	Title = {A review of studies on expert estimation of software development effort},
	Volume = {70},
	Year = {2004}}

@book{Alpaydin2004,
	Author = {Ethem Alpaydin},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Owner = {ekrem},
	Publisher = {MIT Press},
	Timestamp = {2009.10.29},
	Title = {Introduction to Machine Learning},
	Year = {2004}}

@mastersthesis{baker07,
	Author = {Dan Baker},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Note = {Available from \url{https://eidr.wvu.edu/etd/documentdata.eTD?documentid=5443}},
	School = {Lane Department of Computer Science and Electrical Engineering, West Virginia University},
	Title = {A Hybrid Approach to Expert and Model-based Effort Estimation},
	Year = {2007}}

@article{Bakir2009,
	Author = {Bakir, Ayse and Turhan, Burak and Bener, Ayse},
	Citeulike-Article-Id = {5109598},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {10.1007/s11219-009-9081-z},
	Journal = {Software Quality Journal},
	Owner = {ekrem},
	Posted-At = {2009-07-10 18:58:59},
	Timestamp = {2009.10.28},
	Title = {A new perspective on data homogeneity in software cost estimation: A study in the embedded systems domain},
	Url = {http://dx.doi.org/10.1007/s11219-009-9081-z},
	Year = {2009},
	Bdsk-Url-1 = {http://dx.doi.org/10.1007/s11219-009-9081-z}}

@book{Boehm1981,
	Address = {Upper Saddle River, NJ, USA},
	Author = {Boehm, Barry W.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Isbn = {0138221227},
	Publisher = {Prentice Hall PTR},
	Title = {Software Engineering Economics},
	Year = {1981}}

@book{Breimann1984,
	Address = {Monterey, CA},
	Author = {L. Breiman and J. Friedman and R. Olshen and C. Stone},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-19 13:32:06 +1000},
	Owner = {ekrem},
	Publisher = {Wadsworth and Brooks},
	Timestamp = {2010.08.07},
	Title = {Classification and Regression Trees},
	Year = {1984}}

@inproceedings{Briand1999,
	Address = {New York, NY, USA},
	Author = {Briand, Lionel C. and El Emam, Khaled and Surmann, Dagmar and Wieczorek, Isabella and Maxwell, Katrina D.},
	Booktitle = {ICSE '99: Proceedings of the 21st international conference on Software engineering},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://doi.acm.org/10.1145/302405.302647},
	Isbn = {1-58113-074-0},
	Location = {Los Angeles, California, United States},
	Pages = {313--322},
	Publisher = {ACM},
	Title = {An assessment and comparison of common software cost estimation modeling techniques},
	Year = {1999},
	Bdsk-Url-1 = {http://doi.acm.org/10.1145/302405.302647}}

@article{chang74,
	Author = {C.L. Chang},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {IEEE Trans. on Computers},
	Pages = {1179-1185},
	Title = {Finding Prototypes for Nearest Neighbor Classifiers},
	Year = {1974}}

@inproceedings{FayIra93Multi,
	Author = {U M Fayyad and I H Irani},
	Booktitle = {Proceedings of the Thirteenth International Joint Conference on Artificial Intelligence},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Pages = {1022--1027},
	Title = {Multi-interval Discretization of Continuous-valued Attributes for Classification Learning},
	Year = {1993}}

@article{foss03,
	Author = {Foss, T and Stensrud, E and Kitchenham, B and Myrtveit, I},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-19 13:28:43 +1000},
	Journal = {IEEE Transactions on Software Engineering},
	Title = {A simulation study of the model evaluation criterion MMRE},
	Year = {2003}}

@inproceedings{gama06,
	Address = {New York, NY, USA},
	Author = {Joao Gama and Carlos Pinto},
	Booktitle = {SAC '06: Proceedings of the 2006 ACM symposium on Applied computing},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Isbn = {1-59593-108-2},
	Location = {Dijon, France},
	Note = {Available from \url{http://www.liacc.up.pt/~jgama/IWKDDS/Papers/p6.pdf }},
	Pages = {662--667},
	Publisher = {ACM Press},
	Title = {Discretization from data streams: applications to histograms and data mining},
	Year = {2006}}

@article{hall03,
	Author = {M.A. Hall and G. Holmes},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {IEEE Transactions On Knowledge And Data Engineering},
	Number = {6},
	Pages = {1437-1447},
	Title = {Benchmarking Attribute Selection Techniques for Discrete Class Data Mining},
	Volume = {15},
	Year = {2003}}

@article{Hornik1989,
	Address = {Oxford, UK, UK},
	Author = {Hornik, K. and Stinchcombe, M. and White, H.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://dx.doi.org/10.1016/0893-6080(89)90020-8},
	Issn = {0893-6080},
	Journal = {Neural Netw.},
	Number = {5},
	Owner = {ekrem},
	Pages = {359--366},
	Publisher = {Elsevier Science Ltd.},
	Timestamp = {2010.08.07},
	Title = {Multilayer feedforward networks are universal approximators},
	Volume = {2},
	Year = {1989},
	Bdsk-Url-1 = {http://dx.doi.org/10.1016/0893-6080(89)90020-8}}

@article{Kadoda2000,
	Author = {Kadoda, G and Cartwright, M and Shepperd, M},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {UK CBR Workshop, Cambridge, UK},
	Owner = {ekrem},
	Pages = {1--10},
	Timestamp = {2009.10.21},
	Title = {On configuring a case-based reasoning software project prediction system},
	Year = {2000}}
@inproceedings{keung2008a,
	Address = {New York, NY, USA},
	Author = {Keung, Jacky},
	Booktitle = {ESEM '08: Proceedings of the Second ACM-IEEE international symposium on Empirical software engineering and measurement},
	Doi = {http://doi.acm.org/10.1145/1414004.1414057},
	Isbn = {978-1-59593-971-5},
	Location = {Kaiserslautern, Germany},
	Pages = {294--296},
	Publisher = {ACM},
	Title = {Empirical evaluation of analogy-x for software cost estimation},
	Year = {2008},
	Bdsk-Url-1 = {http://doi.acm.org/10.1145/1414004.1414057}}

@inproceedings{keung2008c,
	Address = {Washington, DC, USA},
	Author = {Keung, Jacky and Kitchenham, Barbara},
	Booktitle = {ASWEC '08: Proceedings of the 19th Australian Conference on Software Engineering},
	Isbn = {978-0-7695-3100-7},
	Pages = {229--238},
	Publisher = {IEEE Computer Society},
	Title = {Experiments with Analogy-X for Software Cost Estimation},
	Year = {2008}}

@article{keung2008b,
	Address = {Piscataway, NJ, USA},
	Author = {Keung, Jacky Wai and Kitchenham, Barbara A. and Jeffery, David Ross},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://dx.doi.org/10.1109/TSE.2008.34},
	Issn = {0098-5589},
	Journal = {IEEE Trans. Softw. Eng.},
	Number = {4},
	Pages = {471--484},
	Publisher = {IEEE Press},
	Title = {Analogy-X: Providing Statistical Inference to Analogy-Based Software Cost Estimation},
	Volume = {34},
	Year = {2008},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/TSE.2008.34}}

@article{Kirsopp2002,
	Author = {Kirsopp, C. and Shepperd, M.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {10.1049/ip-sen},
	Issue = {5},
	Journal = {IEEE Proc.},
	Mendeley-Groups = {Automatically Imported},
	Owner = {ekrem},
	Timestamp = {2009.11.26},
	Title = {Making inferences with small numbers of training sets},
	Volume = {149},
	Year = {2002},
	Bdsk-Url-1 = {http://dx.doi.org/10.1049/ip-sen}}

@article{Kirsopp2002a,
	Address = {San Francisco, CA, USA},
	Author = {Kirsopp, Colin and Shepperd, Martin J. and Hart, John},
	Booktitle = {GECCO '02: Proceedings of the Genetic and Evolutionary Computation Conference},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Isbn = {1-55860-878-8},
	Pages = {1367--1374},
	Publisher = {Morgan Kaufmann Publishers Inc.},
	Title = {Search Heuristics, Case-based Reasoning And Software Project Effort Prediction},
	Year = {2002}}

@article{Li2009,
	Author = {Li, Y and Xie, M and Goh, T},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {10.1016/j.jss.2008.06.001},
	Issn = {01641212},
	Issue = {2},
	Journal = {Journal of Systems and Software},
	Pages = {241--252},
	Title = {A study of project selection and feature weighting for analogy based software cost estimation},
	Volume = {82},
	Year = {2009},
	Bdsk-Url-1 = {http://dx.doi.org/10.1016/j.jss.2008.06.001}}

@article{Li2009a,
	Author = {Li, Y. and Xie, M. and Goh T.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {Empirical Software Engineering},
	Pages = {603-643},
	Title = {A study of the non-linear adjustment for analogy based software cost estimation},
	Year = {2009}}

@article{Lipowezky1998,
	Author = {Lipowezky, U.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Issue = {10},
	Journal = {Pattern Recognition Letters},
	Pages = {907-918},
	Title = {Selection of the optimal prototype subset for 1-NN classification},
	Volume = {19},
	Year = {1998}}

@article{Mendes2003,
	Author = {Emilia Mendes and Ian D. Watson and Chris Triggs and Nile Mosley and Steve Counsell},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {Empirical Software Engineering},
	Number = {2},
	Pages = {163-196},
	Title = {A Comparative Study of Cost Estimation Models for Web Hypermedia Applications},
	Volume = {8},
	Year = {2003}}

@article{Mendes2007,
	Address = {Piscataway, NJ, USA},
	Author = {Kitchenham, Barbara and Mendes, Emilia and Travassos, Guilherme H.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://dx.doi.org/10.1109/TSE.2007.1001},
	Issn = {0098-5589},
	Journal = {IEEE Trans. Softw. Eng.},
	Note = {Member-Kitchenham, Barbara A.},
	Number = {5},
	Pages = {316--329},
	Publisher = {IEEE Press},
	Title = {Cross versus Within-Company Cost Estimation Studies: A Systematic Review},
	Volume = {33},
	Year = {2007},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/TSE.2007.1001}}

@article{Menzies2006,
	Author = {Menzies, Tim and Chen, Zhihao and Hihn, Jairus and Lum, Karen},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {10.1109/TSE.2006.114},
	Journal = {IEEE Transactions on Software Engineering},
	Pages = {883--895},
	Title = {Selecting Best Practices for Effort Estimation},
	Volume = {32},
	Year = {2006},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/TSE.2006.114}}

@inproceedings{Milic2004,
	Author = {Milic, Drazen and Wohlin, Claes},
	Booktitle = {Euromicro},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Title = {Distribution Patterns of Effort Estimations},
	Year = {2004}}

@article{Miyazaki1994,
	Address = {New York, NY, USA},
	Author = {Miyazaki, Y. and Terakado, M. and Ozaki, K. and Nozaki, H.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://dx.doi.org/10.1016/0164-1212(94)90110-4},
	Issn = {0164-1212},
	Journal = {J. Syst. Softw.},
	Number = {1},
	Pages = {3--16},
	Publisher = {Elsevier Science Inc.},
	Title = {Robust regression for developing software estimation models},
	Volume = {27},
	Year = {1994},
	Bdsk-Url-1 = {http://dx.doi.org/10.1016/0164-1212(94)90110-4}}

@article{Robson2002,
	Author = {Robson, C},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {Blackwell Publisher Ltd},
	Title = {Real world research: a resource for social scientists and practitioner-researchers},
	Year = {2002}}

@article{Stensrud,
	Address = {Hingham, MA, USA},
	Author = {Stensrud, Erik and Foss, Tron and Kitchenham, Barbara and Myrtveit, Ingunn},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://dx.doi.org/10.1023/A:1023010612345},
	Issn = {1382-3256},
	Journal = {Empirical Softw. Engg.},
	Number = {2},
	Pages = {139--161},
	Publisher = {Kluwer Academic Publishers},
	Title = {A Further Empirical Investigation of the Relationship Between MRE and Project Size},
	Volume = {8},
	Year = {2003},
	Bdsk-Url-1 = {http://dx.doi.org/10.1023/A:1023010612345}}

@article{Walkerden1999,
	Address = {Hingham, MA, USA},
	Author = {Walkerden, Fiona and Jeffery, Ross},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Doi = {http://dx.doi.org/10.1023/A:1009872202035},
	Issn = {1382-3256},
	Journal = {Empirical Softw. Engg.},
	Number = {2},
	Pages = {135--158},
	Publisher = {Kluwer Academic Publishers},
	Title = {An Empirical Study of Analogy-based Software Effort Estimation},
	Volume = {4},
	Year = {1999},
	Bdsk-Url-1 = {http://dx.doi.org/10.1023/A:1009872202035}}

@article{Wang2009,
	Author = {Wang, Y. and Song, Q. and MacDonell, S. and Shepperd, M. and Shen, J.},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Journal = {IEEE Transactions on Systems},
	Owner = {ekrem},
	Pages = {647 - 658},
	Timestamp = {2010.01.14},
	Title = {Integrate the GM(1,1) and Verhulst Models to Predict Software Stage-Effort},
	Volume = {39},
	Year = {2009}}

@inproceedings{YanWeb02Comparative,
	Author = {Ying Yang and Geoffrey I. Webb},
	Booktitle = {Proceedings of PKAW 2002: The 2002 Pacific Rim Knowledge Acquisition Workshop},
	Date-Added = {2010-08-16 09:34:21 +1000},
	Date-Modified = {2010-08-16 09:34:21 +1000},
	Pages = {159-173},
	Title = {A Comparative Study of Discretization Methods forNaive-Bayes Classifiers},
	Year = {2002}}

@inproceedings{Jor2000,
	Abstract = {Reports an empirical study of 109 randomly selected maintenance tasks in a large Norwegian software organization. When the maintainers had understood the maintenance task specifications, we asked whether they knew how to solve the task. A high confidence in knowing how to solve the task meant that the maintainers did not expect any major difficulties. Then, immediately after the task was completed, we asked whether there had been any major unexpected difficulties. A comparison of the answers gave the seemingly surprising result that one could not, except for corrective, small and simple maintenance tasks, have more confidence in the predictions of an experienced maintainer than the predictions of an inexperienced maintainer. We believe that better quality of the feedback on previous predictions and more training in probabilistic thinking are important means to improve the prediction abilities of maintainers. Decision aids, such as maintenance effort estimation models, should enable the analysis of previous predictions and stimulate probabilistic thinking},
	Author = {Jorgensen, M. and Sjoberg, D. I. K. and Kirkeboen, G.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Keywords = {personnel software development management software maintenance training Norwegian software organization confidence decision aids experienced software maintainers feedback quality maintenance effort estimation models prediction ability probabilistic thinking software maintenance tasks task specification training unexpected difficulties},
	Pages = {93-99},
	Title = {The prediction ability of experienced software maintainers},
	Year = {2000}}

@inproceedings{Jor2003,
	Abstract = {This paper presents a process framework and a preliminary checklist for software cost management. While most textbooks and research papers on cost estimation look mainly at the estimation phase, our framework and checklist includes the phases relevant to estimation: preparation, estimation, application, and learning. We believe that cost estimation processes and checklists should support these phases to enable high estimation accuracy. The checklist we suggest is based on checklists from a number of sources, e.g., a handbook in forecasting and checklists present in several Norwegian software companies, it needs, however, to be extended through feedback from other researchers and software practitioners. There is also a need for a provision of conditions for meaningful use of the checklist issues and descriptions of the strength and sources of evidence in favor of the checklist issues. The present version of the checklist should therefore be seen as preliminary and we want to get feedback from the conference participants and other readers of this paper for further improvements.},
	Author = {Jorgensen, M. and Molokken, K.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Keywords = {software cost estimation software quality risk management software cost estimation checklist software cost management software quality assurance software quality management},
	Pages = {134-140},
	Title = {A preliminary checklist for software cost management},
	Year = {2003}}

@article{Jor2004a,
	Abstract = {Traditionally, software professionals are requested to provide minimum-maximum intervals to indicate the uncertainty of their effort estimates. We claim that the traditional request is not optimal and leads to overoptimistic views about the level of estimation uncertainty. Instead, we propose that it is better to frame the request for uncertainty assessment: How likely is it that the actual effort will be more than/less than X? Our claim is based on the results of a previously reported-experiment and field studies in two companies. The two software companies were instructed to apply the traditional and our alternative framing on random samples of their projects. In total, we collected information about 47 projects applying the traditional-framing and 23 projects applying the alternative framing.},
	Author = {Jorgensen, M.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {professional aspects project management risk management software cost estimation software houses alternative framing project risk assessment software companies software cost estimation software professionals software psychology traditional-framing project},
	Note = {0098-5589},
	Number = {4},
	Pages = {209-217},
	Title = {Realism in assessment of effort estimation uncertainty: it matters how you ask},
	Volume = {30},
	Year = {2004}}

@inproceedings{Jor2004b,
	Abstract = {This paper describes the mechanical and electrical design of a new lattice based self-reconfigurable robot, called the ATRON. The ATRON system consists of several fully self-contained robot modules, each having their own processing power, power supply, sensors and actuators. The ATRON modules are roughly spheres with equatorial rotation. Each module can be connected to up to eight neighbors through four male and four female connectors. In this paper, we describe the realization of the design, both the mechanics and the electronics. Details on power sharing and power consumption is given. Finally, this paper includes a brief outline of our future work on the ATRON system.},
	Author = {Jorgensen, M. W. and Ostergaard, E. H. and Lund, H. H.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Keywords = {control system synthesis robots self-adjusting systems modular ATRON power consumption power sharing self-contained robot modules self-reconfigurable robot},
	Pages = {2068-2073 vol.2},
	Title = {Modular ATRON: modules for a self-reconfigurable robot},
	Volume = {2},
	Year = {2004}}

@article{Jor2004c,
	Abstract = {This study aims to improve analyses of why errors occur in software effort estimation. Within one software development company, we collected information about estimation errors through: 1) interviews with employees in different roles who are responsible for estimation, 2) estimation experience reports from 68 completed projects, and 3) statistical analysis of relations between characteristics of the 68 completed projects and estimation error. We found that the role of the respondents, the data collection approach, and the type of analysis had an important impact on the reasons given for estimation error. We found, for example, a strong tendency to perceive factors outside the respondents' own control as important reasons for inaccurate estimates. Reasons given for accurate estimates, on the other hand, typically cited factors that were within the respondents' own control and were determined by the estimators' skill or experience. This bias in types of reason means that the collection only of project managers' viewpoints will not yield balanced models of reasons for estimation error. Unfortunately, previous studies on reasons for estimation error have tended to collect information from project managers only. We recommend that software companies combine estimation error information from in-depth interviews with stakeholders in all relevant roles, estimation experience reports, and results from statistical analyses of project characteristics.},
	Author = {Jorgensen, M. and Molokken-Ostvold, K.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {cost-benefit analysis error handling project management software cost estimation software development management software performance evaluation cost estimation data analysis information collection approach interviews performance evaluation project evaluation project management software development software effort estimation error software review statistical analysis},
	Note = {0098-5589},
	Number = {12},
	Pages = {993-1007},
	Title = {Reasons for software effort estimation error: impact of respondent role, information collection approach, and data analysis method},
	Volume = {30},
	Year = {2004}}

@article{Jor2004d,
	Abstract = {The study described in this paper reports from a real-life bidding process in which 35 companies were bidding for the same contract. The bidding process consisted of two separate phases: a prestudy phase and a bidding phase. In the prestudy phase, 17 of the 35 bidding companies provided rough price indications based on a brief, incomplete description of user requirements. In the bidding phase, all 35 companies provided bids based on a more complete requirement specification that described a software system with substantially more functionality than the system indicated in the prestudy phase. The main result of the study is that the 17 companies involved in the prestudy phase presented bids that were, on average, about 70 percent higher than the bids of the other companies, although all companies based their bids on the same requirement specification. We propose an explanation for this difference that is consistent with the prospect theory and the precautionary bidding effect. A possible implication of our findings is that software clients should not request early price indications based on limited and uncertain information when the final bids can be based on more complete and reliable information.},
	Author = {Jorgensen, M. and Carelius, G. J.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {DP industry contracts formal specification project management risk management software cost estimation software management bidding phase prestudy phase requirement specification software client software project bidding software system},
	Note = {0098-5589},
	Number = {12},
	Pages = {953-969},
	Title = {An empirical study of software project bidding},
	Volume = {30},
	Year = {2004}}

@inproceedings{Jor2005,
	Abstract = {Evidence-based software engineering (EBSE) describes a process of identifying, understanding and evaluating findings from research and practice-based experience. This process aims at improving software engineering decisions. For the last three years, EBSE has been taught to university students at Hedmark University College, Rena, Norway. The motivation for the EBSE-course is that it is essential for the students, as future practitioners, to learn how to base important software engineering decisions on the systematic and critical evaluation of the best available evidence. The main purpose of this paper is to inspire and support other universities in their work on developing their own EBSE-courses. For this purpose we report on how our course has been organized and what lessons have been learned. There are currently no studies available on the effects of teaching EBSE and, as far as we know, only we have gained practice-based experience. To acquire more knowledge about the costs and benefits of teaching EBSE we hope that other universities will develop their own EBSE-courses and report their experience.},
	Author = {Jorgensen, M. and Dyba, T. and Kitchenham, B.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Pages = {24-24},
	Title = {Teaching Evidence-Based Software Engineering to University Students},
	Year = {2005}}

@article{Jor2005a,
	Abstract = {Several studies suggest that uncertainty assessments of software development costs are strongly biased toward overconfidence, i.e., that software cost estimates typically are believed to be more accurate than they really are. This overconfidence may lead to poor project planning. As a means of improving cost uncertainty assessments, we provide evidence-based guidelines for how to assess software development cost uncertainty, based on results from relevant empirical studies. The general guidelines provided are: 1) Do not rely solely on unaided, intuition-based uncertainty assessment processes, 2) do not replace expert judgment with formal uncertainty assessment models, 3) apply structured and explicit judgment-based processes, 4) apply strategies based on an outside view of the project, 5) combine uncertainty assessments from different sources through group work, not through mechanical combination, 6) use motivational mechanisms with care and only if greater effort is likely to lead to improved assessments, and 7) frame the assessment problem to fit the structure of the relevant uncertainty information and the assessment process. These guidelines are preliminary and should be updated in response to new evidence.},
	Author = {Jorgensen, M.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {formal specification project management software cost estimation evidence-based guideline formal uncertainty assessment model intuition-based uncertainty assessment mechanical combination project planning software cost estimation software development cost uncertainty software psychology Index Terms- Cost estimation management software psychology uncertainty of software development cost.},
	Note = {0098-5589},
	Number = {11},
	Pages = {942-954},
	Title = {Evidence-based guidelines for assessment of software development cost uncertainty},
	Volume = {31},
	Year = {2005}}

@article{Jor2005b,
	Abstract = {This article presents seven guidelines for producing realistic software development effort estimates. The guidelines derive from industrial experience and empirical studies. While many other guidelines exist for software effort estimation, these guidelines differ from them in three ways: 1) They base estimates on expert judgments rather than models. 2) They are easy to implement. 3) They use the most recent findings regarding judgment-based effort estimation. Estimating effort on the basis of expert judgment is the most common approach today, and the decision to use such processes instead of formal estimation models shouldn't be surprising. Simple process changes such as reframing questions can lead to more realistic estimates of software development efforts.},
	Author = {Jorgensen, M.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Journal = {Software, IEEE},
	Keywords = {project management software cost estimation software development management expert judgment approach project management software development software effort estimation project management software cost estimation software development management},
	Note = {0740-7459},
	Number = {3},
	Pages = {57-63},
	Title = {Practical guidelines for expert-judgment-based software effort estimation},
	Volume = {22},
	Year = {2005}}

@inproceedings{Jor2005c,
	Abstract = {It is well known that software development projects tend to be based on over-optimistic cost estimates. Better knowledge of the sources of this over-optimism is necessary to improve realism in software development project bids and budgets. This paper analyses the effect of the winner&\#146;s curse. The winner&\#146;s curse is a result of the selection of software providers among those with the lowest bid, i.e., those with a tendency towards the highest level of over-optimism. The winner&\#146;s curse has not been extensively analyzed in software cost estimation studies, but is a well known phenomenon in domains such as auctioning. We exemplify the effect of the winner&\#146;s curse with data from a real software development bidding round and simulate how increase in number of bidders and cost uncertainty impact the expected profit. We argue that the winners&\#146; curse is a problem for both clients and providers, and that it may lead to inefficient use of scarce resources. Possible remedies for reduction of negative effects of the winner&\#146;s curse are discussed.},
	Author = {Jorgensen, M. and Grimstad, S.},
	Date-Added = {2010-08-15 20:37:06 +1000},
	Date-Modified = {2010-08-15 20:37:06 +1000},
	Pages = {280-285},
	Title = {Over-Optimism in Software Development Projects: The Winner&\#146;s Curse},
	Year = {2005}}

@article{kleijnen97,
	Author = {J.P.C. Kliijnen},
	Date-Added = {2010-08-15 14:42:20 +1000},
	Date-Modified = {2010-08-15 14:42:20 +1000},
	Journal = {Journal Statistical Computation and Simulation},
	Number = {1--4},
	Pages = {111-142},
	Title = {Sensitivity Analysis and Related Analyses: a Survey of Statistical Techniques},
	Volume = 57,
	Year = {1997}}

@article{kitchenham_07,
	Author = {Kitchenham, B.A. and Mendes, E. and Travassos, G.H.},
	Date-Modified = {2010-08-15 12:41:16 +1000},
	Doi = {10.1109/TSE.2007.1001},
	Issn = {0098-5589},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {cross-company-based estimation;software cost estimation models;software engineering;software cost estimation;},
	Month = {may},
	Number = {5},
	Pages = {316 -329},
	Title = {Cross versus Within-Company Cost Estimation Studies: A Systematic Review},
	Volume = {33},
	Year = {2007},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/TSE.2007.1001}}

@inproceedings{mendes_04,
	Address = {Washington, DC, USA},
	Author = {Mendes, Emilia and Kitchenham, Barbara},
	Booktitle = {METRICS '04: Proceedings of the Software Metrics, 10th International Symposium},
	Date-Modified = {2010-08-15 12:40:33 +1000},
	Doi = {http://dx.doi.org/10.1109/METRICS.2004.24},
	Isbn = {0-7695-2129-0},
	Pages = {348--357},
	Publisher = {IEEE Computer Society},
	Title = {Further Comparison of Cross-Company and Within-Company Effort Estimation Models for Web Applications},
	Year = {2004},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRICS.2004.24}}

@inproceedings{mendes_05,
	Address = {Washington, DC, USA},
	Author = {Mendes, Emilia and Lokan, Chris and Harrison, Robert and Triggs, Chris},
	Booktitle = {METRICS '05: Proceedings of the 11th IEEE International Software Metrics Symposium},
	Date-Modified = {2010-08-15 12:41:05 +1000},
	Doi = {http://dx.doi.org/10.1109/METRICS.2005.4},
	Isbn = {0-7695-2371-4},
	Pages = {36},
	Publisher = {IEEE Computer Society},
	Title = {A Replicated Comparison of Cross-Company and Within-Company Effort Estimation Models Using the ISBSG Database},
	Year = {2005},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRICS.2005.4}}

@inproceedings{briand99,
	Abstract = {The use of test coverage measures (e.g. block coverage) to control the software test process has become an increasingly common practice. This is justified by the assumption that higher test coverage helps achieve higher defect coverage and therefore improves software quality. In practice, data often shows that defect coverage and test coverage grow over time, as additional testing is performed. However, it is unclear whether this phenomenon of concurrent growth can be attributed to a causal dependency or if it is coincidental, simply due to the cumulative nature of both measures. Answering such a question is important as it determines whether a given test coverage measure should be monitored for quality control and used to drive testing. Although this is no general answer to the problem above, we propose a procedure to investigate whether any test coverage criterion has a genuine additional impact on defect coverage when compared to the impact of just running additional test cases. This procedure is applicable in typical testing conditions where the software is tested once, according to a given strategy and where coverage measures are collected as well as defect data. We then test the procedure on published data and compare our results with the original findings. The study outcomes do not support the assumption of a causal dependency between test coverage and defect coverage, a result for which several plausible explanations are provided},
	Author = {Briand, L. and Pfahl, D.},
	Booktitle = {Software Reliability Engineering, 1999. Proceedings. 10th International Symposium on},
	Date-Modified = {2010-08-03 16:35:00 +1000},
	Doi = {10.1109/ISSRE.1999.809319},
	Keywords = {causal dependency;defect coverage;quality control;simulation;software quality;software testing;test coverage;program testing;software quality;},
	Pages = {148 -157},
	Title = {Using simulation for assessing the real impact of test coverage on defect coverage},
	Year = {1999},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ISSRE.1999.809319}}

@inproceedings{keung08b,
	Abstract = {Software cost estimation is an important area of research in software engineering. Various cost estimation model evaluation criteria (such as MMRE, MdMRE etc.) have been developed for comparing prediction accuracy among cost estimation models. All of these metrics capture the residual difference between the predicted value and the actual value in the dataset, but ignore the importance of the dataset quality. What is more, they implicitly assume the prediction model to be able to predict with up to 100% accuracy at its maximum for a given dataset. Given that these prediction models only provide an estimate based on observed historical data, absolute accuracy cannot be possibly achieved. It is therefore important to realize the theoretical maximum prediction accuracy (TMPA) for the given model with a given dataset. In this paper, we first discuss the practical importance of this notion, and propose a novel method for the determination of TMPA in the application of analogy-based software cost estimation. Specifically, we determine the TMPA of analogy using a unique dynamic K-NN approach to simulate and optimize the prediction system. The results of an empirical experiment show that our method is practical and important for researchers seeking to develop improved prediction models, because it offers an alternative for practical comparison between different prediction models.},
	Author = {Keung, J.W.},
	Booktitle = {Software Engineering Conference, 2008. APSEC '08. 15th Asia-Pacific},
	Date-Modified = {2010-08-03 16:37:20 +1000},
	Doi = {10.1109/APSEC.2008.43},
	Issn = {1530-1362},
	Keywords = {analogy-based software cost estimation;prediction models;software metrics;theoretical maximum prediction accuracy;software cost estimation;software metrics;},
	Month = {3-5},
	Pages = {495 -502},
	Title = {Theoretical Maximum Prediction Accuracy for Analogy-Based Software Cost Estimation},
	Year = {2008},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/APSEC.2008.43}}

@inproceedings{stensrud02,
	Abstract = { Cost estimates are important deliverables of a software project. Consequently, a number of cost prediction models have been proposed and evaluated. The common evaluation criteria have been MMRE, MdMRE and PRED(k). MRE is the basic metric in these evaluation criteria. The implicit rationale of using a relative error measure like MRE, rather than an absolute one, is presumably to have a measure that is independent of project size. We investigate if this implicit claim holds true for several data sets: Albrecht, Kemerer, Finnish, DMR and Accenture-ERP. The results suggest that MRE is not independent of project size. Rather, MRE is larger for small projects than for large projects. A practical consequence is that a project manager predicting a small project may falsely believe in a too low MRE. Vice versa when predicting a large project. For researchers, it is important to know that MMRE is not an appropriate measure of the expected MRE of small and large projects. We recommend therefore that the data set be partitioned into two or more subsamples and that MMRE is reported per subsample. In the long term, we should consider using other evaluation criteria.},
	Author = {Stensrud, E. and Foss, T. and Kitchenham, B. and Myrtveit, I.},
	Booktitle = {Software Metrics, 2002. Proceedings. Eighth IEEE Symposium on},
	Date-Modified = {2010-08-03 16:35:46 +1000},
	Doi = {10.1109/METRIC.2002.1011320},
	Keywords = {Accenture-ERP data set; Albrecht data set; DMR data set; Finnish data set; Kemerer data set; MMRE; MRE; MdMRE; PRED; common evaluation criteria; cost prediction models; magnitude of relative error; relative error measure; software cost estimates; software metrics; software project size; project management; software cost estimation; software development management; software metrics;},
	Pages = {3 - 12},
	Title = {An empirical validation of the relationship between the magnitude of relative error and project size},
	Year = {2002},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRIC.2002.1011320}}

@article{kitchenham01,
	Abstract = {Provides the software estimation research community with a better understanding of the meaning of, and relationship between, two statistics that are often used to assess the accuracy of predictive models: the mean magnitude relative error (MMRE) and the number of predictions within 25% of the actual, pred(25). It is demonstrated that MMRE and pred(25) are, respectively, measures of the spread and the kurtosis of the variable z, where z=estimate/actual. Thus, z is considered to be a measure of accuracy, and statistics such as MMRE and pred(25) to be measures of properties of the distribution of z. It is suggested that measures of the central location and skewness of z, as well as measures of spread and kurtosis, are necessary. Furthermore, since the distribution of z is non-normal, non-parametric measures of these properties may be needed. For this reason, box-plots of z are useful alternatives to simple summary metrics. It is also noted that the simple residuals are better behaved than the z variable, and could also be used as the basis for comparing prediction systems},
	Author = {Kitchenham, B.A. and Pickard, L.M. and MacDonell, S.G. and Shepperd, M.J.},
	Date-Modified = {2010-08-03 16:35:29 +1000},
	Doi = {10.1049/ip-sen:20010506},
	Issn = {1462-5970},
	Journal = {Software, IEE Proceedings -},
	Keywords = {accuracy measures;accuracy statistics;box-plots;central location measure;kurtosis measure;mean magnitude relative error;nonnormal distribution;nonparametric measures;prediction number;prediction systems comparison;predictive models;residuals;skewness measure;software estimation;spread measure;statistical distribution properties;summary metrics;nonparametric statistics;software cost estimation;software metrics;},
	Month = {jun},
	Number = {3},
	Pages = {81 -85},
	Title = {What accuracy statistics really measure [software estimation]},
	Volume = {148},
	Year = {2001},
	Bdsk-Url-1 = {http://dx.doi.org/10.1049/ip-sen:20010506}}

@inproceedings{ruhe03,
	Abstract = {In this paper, we investigate the application of the COBRA trade; method (Cost Estimation, Benchmarking, and Risk Assessment) in a new application domain, the area of web development. COBRA combines expert knowledge with data on a small number of projects to develop cost estimation models, which can also be used for risk analysis and benchmarking purposes. We modified and applied the method to the web applications of a small Australian company, specializing in web development. In this paper we present the modifications made to the COBRA method and results of applying the method In our study, using data on twelve web applications, the estimates derived from our Web-COBRA model showed a Mean Magnitude of Relative Error (MMRE) of 0.17. This result significantly outperformed expert estimates from Allette Systems (MMRE 0.37). A result comparable to Web-COBRA was obtained when applying ordinary least squares regression with size in terms of Web Objects as an independent variable (MMRE 0.23).},
	Author = {Ruhe, M. and Jeffery, R. and Wieczorek, I.},
	Booktitle = {Software Engineering, 2003. Proceedings. 25th International Conference on},
	Date-Modified = {2010-08-03 16:36:20 +1000},
	Doi = {10.1109/ICSE.2003.1201208},
	Issn = {0270-5257},
	Keywords = {COBRA application; Web applications; Web-COBRA model; least squares regression; mean magnitude of relative error; software cost estimation; Internet; distributed object management; regression analysis; software cost estimation;},
	Month = {3-10},
	Pages = {285 - 294},
	Title = {Cost estimation for web applications},
	Year = {2003},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ICSE.2003.1201208}}

@inproceedings{martin05,
	Abstract = { Software estimation has been identified as one of the three great challenges for half-century-old computer science. Developers should be able to achieve practices containing effort estimation based on their own programs. New paradigms as fuzzy logic may offer an alternative for software effort estimation. This paper describes an application whose results are compared with those of a multiple regression. A subset of 41 modules developed from ten programs are used as data. Result shows that the value of MMRE (an aggregation of magnitude of relative error, MRE) applying fuzzy logic was slightly higher than MMRE applying multiple regression; while the value of Pred(20) applying fuzzy logic was slightly higher than Pred(20) applying multiple regression. Moreover, six of 41 MRE was equal to zero (without any deviation) when fuzzy logic was applied (not any similar case was presented when multiple regression was applied).},
	Author = {Martin, C.L. and Pasquier, J.L. and Yanez, C.M. and Tornes, A.G.},
	Booktitle = {Computer Science, 2005. ENC 2005. Sixth Mexican International Conference on},
	Date-Modified = {2010-08-03 16:36:41 +1000},
	Doi = {10.1109/ENC.2005.47},
	Issn = {1550-4069},
	Keywords = {fuzzy logic; multiple regression; relative error; software development; software effort estimation; fuzzy logic; regression analysis; software metrics;},
	Month = {26-30},
	Pages = {113 - 120},
	Title = {Software development effort estimation using fuzzy logic: a case study},
	Year = {2005},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ENC.2005.47}}

@inproceedings{macdonell97,
	Abstract = {An important task for any software project manager is to be able to predict and control project size and development effort. Unfortunately, there is comparatively little work, other than function points, that tackles the problem of building prediction systems for software that is dominated by data considerations, in particular systems developed using 4GLs. We describe an empirical investigation of 70 such systems. Various easily obtainable counts were extracted from data models (e.g. number of entities) and from specifications (e.g. number of screens). Using simple regression analysis, a prediction system of implementation size with accuracy of MMRE=21% was constructed. This approach offers several advantages. First there tend to be fewer counting problems than with function points since the metrics we used were based upon simple counts. Second, the prediction systems were calibrated to specific local environments rather than being based upon industry weights. We believe this enhanced their accuracy. Our work shows that it is possible to develop simple and useful local prediction systems based upon metrics easily derived from functional specifications and data models, without recourse to overly complex metrics or analysis techniques. We conclude that this type of use of metrics can provide valuable support for the management and control of 4GL and database projects},
	Author = {MacDonell, S.G. and Shepperd, M.J. and Sallis, P.J.},
	Booktitle = {Software Metrics Symposium, 1997. Proceedings., Fourth International},
	Date-Modified = {2010-08-03 16:34:44 +1000},
	Doi = {10.1109/METRIC.1997.637170},
	Keywords = {4GLs;counting problems;data considerations;data models;database project management;database system metrics;empirical study;function points;functional specifications;implementation size;local prediction systems;prediction system;prediction systems;project size;simple regression analysis;software project manager;data structures;formal specification;project management;software metrics;software reliability;statistical analysis;},
	Month = {5-7},
	Pages = {99 -107},
	Title = {Metrics for database systems: an empirical study},
	Year = {1997},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRIC.1997.637170}}

@inproceedings{wu06,
	Abstract = {Over the past ten couple of years, there is a variety of effort models proposed by academicians and practitioners at early stage of software development life cycle. Some addressed that efforts could be predicted using lines of codes (LOC) and COCOMO, others emphasized that it could be made using function point analysis (FPA) or others. The study seeks to develop a model that estimates software effort by studying and analyzing small and medium scale application software. To develop such a model, 50 completed software projects are collected from a software company. With the sample data, design team factors are identified and extracted. By applying them to simple regression analyses, a prediction of software of effort estimates with accuracy of MMRE=9% was constructed. The results give several benefits. First, the estimation problems are minimized due to the simple procedure used in identifying those factors. Second, the predicted software projects are only limited to a specific environment rather than being based upon industry environment. We believe the accuracy of effort estimates can be improved. According to the results analyzed, the work shows that it is possible to build up simple and useful prediction model based on data extracted at the early stage of software development life cycle. We hope this model can provide valuable ideas and suggestions for project designers for planning and controlling software projects in near future},
	Author = {Wu, S.I.K.},
	Booktitle = {Service Operations and Logistics, and Informatics, 2006. SOLI '06. IEEE International Conference on},
	Date-Modified = {2010-08-03 16:36:57 +1000},
	Doi = {10.1109/SOLI.2006.328973},
	Keywords = {application software;design team factor;regression analysis;software development life cycle;software effort estimation;software project;software quality;project management;regression analysis;software cost estimation;software development management;software quality;},
	Month = {21-23},
	Pages = {6 -11},
	Title = {The quality of design team factors on software effort estimation},
	Year = {2006},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/SOLI.2006.328973}}

@inproceedings{dimartino07,
	Abstract = {Size represents one of the most important attribute of software products used to predict software development effort. In the past nine years, several measures have been proposed to estimate the size of Web applications, and it is important to determine which one is most effective to predict Web development effort. To this aim in this paper we report on an empirical analysis where, using data from 15 Web projects developed by a software company, we compare four sets of size measures, using two prediction techniques, namely Forward Stepwise Regression (SWR) and Case-Based Reasoning (CBR). All the measures provided good predictions in terms of MMRE, MdMRE, and Pred(0.25) statistics, for both SWR and CBR. Moreover, when using SWR, length measures and Web Objects gave significant better results than Functional measures, however presented similar results to the Tukutuku measures. As for CBR, results did not show any significant differences amongst the four sets of size measures.},
	Author = {Di Martino, S. and Ferrucci, F. and Gravino, C. and Mendes, E.},
	Booktitle = {Empirical Software Engineering and Measurement, 2007. ESEM 2007. First International Symposium on},
	Date-Modified = {2010-08-03 16:37:07 +1000},
	Doi = {10.1109/ESEM.2007.20},
	Issn = {1938-6451},
	Keywords = {CBR technique;SWR technique;Web application development effort prediction;case-based reasoning;forward stepwise regression;size measures;software products;Web design;case-based reasoning;regression analysis;software metrics;},
	Month = {20-21},
	Pages = {324 -333},
	Title = {Comparing Size Measures for Predicting Web Application Development Effort: A Case Study},
	Year = {2007},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ESEM.2007.20}}

@inproceedings{mendes02b,
	Abstract = { To date studies using CBR for Web hypermedia effort prediction have not applied adaptation rules to adjust effort according to a given criterion. In addition, when applying n-fold cross-validation, their analysis has been limited to a maximum of three training sets, which according to recent studies, may lead to untrustworthy results. This paper has therefore two objectives. The first is to further investigate the use of CBR for Web hypermedia effort prediction by comparing the prediction accuracy of eight CBR techniques, of which three have previously been compared. The second objective is to compare the prediction accuracy of the best CBR technique against stepwise regression, using a twenty-fold cross-validation. All prediction accuracies were measured using Mean Magnitude of Relative Error (MMRE), Median Magnitude of Relative Error, Prediction at level 1 (1=25%), and boxplots of the residuals. One dataset was used in the estimation process and, according to all measures of prediction accuracy, stepwise regression showed the best prediction accuracy.},
	Author = {Mendes, E. and Mosley, N.},
	Booktitle = {Empirical Software Engineering, 2002. Proceedings. 2002 International Symposium n},
	Date-Modified = {2010-08-03 16:36:08 +1000},
	Doi = {10.1109/ISESE.2002.1166928},
	Keywords = {Web hypermedia; case based reasoning; dataset; development effort; mean magnitude of relative error; median magnitude of relative error; n-fold cross-validation; prediction accuracy; stepwise regression; Internet; case-based reasoning; hypermedia; software engineering;},
	Pages = {79 - 90},
	Title = {Further investigation into the use of CBR and stepwise regression to predict development effort for Web hypermedia applications},
	Year = {2002},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ISESE.2002.1166928}}

@inproceedings{mendes02a,
	Abstract = { Several studies have compared the prediction accuracy of different types of techniques with emphasis placed on linear and stepwise regressions, and case-based reasoning (CBR). We believe the use of only one type of CBR technique may bias the results, as there are others that can also be used for effort prediction. This paper has two objectives. The first is to compare the prediction accuracy of three CBR techniques to estimate the effort to develop Web hypermedia applications. The second objective is to compare the prediction accuracy of the best CBR technique, according to our findings, against three commonly used prediction models, namely multiple linear regression, stepwise regression and regression trees. One dataset was used in the estimation process and the results showed that different measures of prediction accuracy gave different results. MMRE and MdMRE showed better prediction accuracy for multiple regression models whereas box plots showed better accuracy for CBR.},
	Author = {Mendes, E. and Watson, I. and Triggs, C. and Mosley, N. and Counsell, S.},
	Booktitle = {Software Metrics, 2002. Proceedings. Eighth IEEE Symposium on},
	Date-Modified = {2010-08-03 16:36:14 +1000},
	Doi = {10.1109/METRIC.2002.1011332},
	Keywords = {case-based reasoning; effort prediction models; multiple regression models; prediction accuracy; project management; software projects; web hypermedia; case-based reasoning; hypermedia; project management; software cost estimation; software development management;},
	Pages = {131 - 140},
	Title = {A comparison of development effort estimation techniques for Web hypermedia applications},
	Year = {2002},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRIC.2002.1011332}}

@article{shepperd01b,
	Abstract = {The need for accurate software prediction systems increases as software becomes much larger and more complex. We believe that the underlying characteristics: size, number of features, type of distribution, etc., of the data set influence the choice of the prediction system to be used. For this reason, we would like to control the characteristics of such data sets in order to systematically explore the relationship between accuracy, choice of prediction system, and data set characteristic. It would also be useful to have a large validation data set. Our solution is to simulate data allowing both control and the possibility of large (1000) validation cases. The authors compare four prediction techniques: regression, rule induction, nearest neighbor (a form of case-based reasoning), and neural nets. The results suggest that there are significant differences depending upon the characteristics of the data set. Consequently, researchers should consider prediction context when evaluating competing prediction systems. We observed that the more "messy" the data and the more complex the relationship with the dependent variable, the more variability in the results. In the more complex cases, we observed significantly different results depending upon the particular training set that has been sampled from the underlying data set. However, our most important result is that it is more fruitful to ask which is the best prediction system in a particular context rather than which is the "best" prediction system},
	Author = {Shepperd, M. and Kadoda, G.},
	Date-Modified = {2010-08-03 16:35:38 +1000},
	Doi = {10.1109/32.965341},
	Issn = {0098-5589},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {case-based reasoning;data set characteristics;machine learning;nearest neighbor;neural nets;prediction problem;regression;rule induction;simulation;small data sets;software prediction systems;software prediction technique comparison;training set;case-based reasoning;learning (artificial intelligence);neural nets;software metrics;virtual machines;},
	Month = {nov},
	Number = {11},
	Pages = {1014 -1022},
	Title = {Comparing software prediction techniques using simulation},
	Volume = {27},
	Year = {2001},
	Bdsk-File-1 = {YnBsaXN0MDDUAQIDBAUIJidUJHRvcFgkb2JqZWN0c1gkdmVyc2lvblkkYXJjaGl2ZXLRBgdUcm9vdIABqAkKFRYXGyIjVSRudWxs0wsMDQ4RFFpOUy5vYmplY3RzV05TLmtleXNWJGNsYXNzog8QgASABqISE4ACgAOAB1lhbGlhc0RhdGFccmVsYXRpdmVQYXRo0hgNGRpXTlMuZGF0YU8RAagAAAAAAagAAgAADE1hY2ludG9zaCBIRAAAAAAAAAAAAAAAAAAAAMgHQrtIKwAAABIP7BZTaGVwcGVyZF9UU0VfMjAwMUEucGRmAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEhATyHxZDQAAAAAAAAAAAAIABQAACSAAAAAAAAAAAAAAAAAAAAAGUGFwZXJzABAACAAAyAa2GwAAABEACAAAyHvMbQAAAAEAEAASD+wACEnkAAhJ1AAAkj8AAgBATWFjaW50b3NoIEhEOlVzZXJzOmphY2t5OkRvd25sb2FkczpQYXBlcnM6U2hlcHBlcmRfVFNFXzIwMDFBLnBkZgAOAC4AFgBTAGgAZQBwAHAAZQByAGQAXwBUAFMARQBfADIAMAAwADEAQQAuAHAAZABmAA8AGgAMAE0AYQBjAGkAbgB0AG8AcwBoACAASABEABIAM1VzZXJzL2phY2t5L0Rvd25sb2Fkcy9QYXBlcnMvU2hlcHBlcmRfVFNFXzIwMDFBLnBkZgAAEwABLwAAFQACAAz//wAAgAXSHB0eH1gkY2xhc3Nlc1okY2xhc3NuYW1lox8gIV1OU011dGFibGVEYXRhVk5TRGF0YVhOU09iamVjdF8QOS4uLy4uL1VzZXJzL2phY2t5L0Rvd25sb2Fkcy9QYXBlcnMvU2hlcHBlcmRfVFNFXzIwMDFBLnBkZtIcHSQloiUhXE5TRGljdGlvbmFyeRIAAYagXxAPTlNLZXllZEFyY2hpdmVyAAgAEQAWAB8AKAAyADUAOgA8AEUASwBSAF0AZQBsAG8AcQBzAHYAeAB6AHwAhgCTAJgAoAJMAk4CUwJcAmcCawJ5AoACiQLFAsoCzQLaAt8AAAAAAAACAQAAAAAAAAAoAAAAAAAAAAAAAAAAAAAC8Q==},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/32.965341}}

@inproceedings{shepperd01a,
	Abstract = {The need for accurate software prediction systems increases as software becomes larger and more complex. A variety of techniques have been proposed, but none has proved consistently accurate. The underlying characteristics of the data set influence the choice of the prediction system to be used. It has proved difficult to obtain significant results over small data sets; consequently, we required large validation data sets. Moreover, we wished to control the characteristics of such data sets in order to systematically explore the relationship between accuracy, choice of prediction system and data set characteristics. Our solution has been to simulate data, allowing both control and the possibility of large validation cases. We compared regression, rule induction and nearest neighbours (a form of case-based reasoning). The results suggest that there are significant differences depending upon the characteristics of the data set. Consequently, researchers should consider the prediction context when evaluating competing prediction systems. We also observed that the more ldquo;messy rdquo; the data and the more complex the relationship with the dependent variable, the more variability in the results. This became apparent since we sampled two different training sets from each simulated population of data. In the more complex cases, we observed significantly different results depending upon the training set. This suggests that researchers will need to exercise caution when comparing different approaches and utilise procedures such as bootstrapping in order to generate multiple samples for training purposes},
	Author = {Shepperd, M. and Kadoda, G.},
	Booktitle = {Software Metrics Symposium, 2001. METRICS 2001. Proceedings. Seventh International},
	Date-Modified = {2010-08-03 16:35:23 +1000},
	Doi = {10.1109/METRIC.2001.915542},
	Keywords = {accuracy;bootstrapping;case-based reasoning;data set characteristics;dependent variable relationship;large validation data sets;multiple sample generation;nearest neighbours;prediction context;regression;results variability;rule induction;simulation;software prediction techniques evaluation;training sets;computer aided software engineering;forecasting theory;inference mechanisms;software metrics;statistical analysis;virtual machines;},
	Pages = {349 -359},
	Title = {Using simulation to evaluate prediction techniques [for software]},
	Year = {2001},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRIC.2001.915542}}

@article{shepperd97,
	Abstract = {Accurate project effort prediction is an important goal for the software engineering community. To date most work has focused upon building algorithmic models of effort, for example COCOMO. These can be calibrated to local environments. We describe an alternative approach to estimation based upon the use of analogies. The underlying principle is to characterize projects in terms of features (for example, the number of interfaces, the development method or the size of the functional requirements document). Completed projects are stored and then the problem becomes one of finding the most similar projects to the one for which a prediction is required. Similarity is defined as Euclidean distance in n-dimensional space where n is the number of project features. Each dimension is standardized so all dimensions have equal weight. The known effort values of the nearest neighbors to the new project are then used as the basis for the prediction. The process is automated using a PC-based tool known as ANGEL. The method is validated on nine different industrial datasets (a total of 275 projects) and in all cases analogy outperforms algorithmic models based upon stepwise regression. From this work we argue that estimation by analogy is a viable technique that, at the very least, can be used by project managers to complement current estimation techniques},
	Author = {Shepperd, M. and Schofield, C.},
	Date-Modified = {2010-08-03 16:34:53 +1000},
	Doi = {10.1109/32.637387},
	Issn = {0098-5589},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {ANGEL;COCOMO;Euclidean distance;algorithmic models;estimation by analogy;functional requirements document;industrial datasets;nearest neighbors;personal computer-based tool;project effort prediction;project management;software development method;software engineering;software project effort estimation;stepwise regression;project management;software cost estimation;software development management;software metrics;software tools;},
	Month = {nov},
	Number = {11},
	Pages = {736 -743},
	Title = {Estimating software project effort using analogies},
	Volume = {23},
	Year = {1997},
	Bdsk-File-1 = {YnBsaXN0MDDUAQIDBAUIJidUJHRvcFgkb2JqZWN0c1gkdmVyc2lvblkkYXJjaGl2ZXLRBgdUcm9vdIABqAkKFRYXGyIjVSRudWxs0wsMDQ4RFFpOUy5vYmplY3RzV05TLmtleXNWJGNsYXNzog8QgASABqISE4ACgAOAB1lhbGlhc0RhdGFccmVsYXRpdmVQYXRo0hgNGRpXTlMuZGF0YU8RAZwAAAAAAZwAAgAADE1hY2ludG9zaCBIRAAAAAAAAAAAAAAAAAAAAMgHQrtIKwAAABIP7BNTaGVwcGVyZF9UU0VfOTcucGRmAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEhCAyHxZwgAAAAAAAAAAAAIABQAACSAAAAAAAAAAAAAAAAAAAAAGUGFwZXJzABAACAAAyAa2GwAAABEACAAAyHvNIgAAAAEAEAASD+wACEnkAAhJ1AAAkj8AAgA9TWFjaW50b3NoIEhEOlVzZXJzOmphY2t5OkRvd25sb2FkczpQYXBlcnM6U2hlcHBlcmRfVFNFXzk3LnBkZgAADgAoABMAUwBoAGUAcABwAGUAcgBkAF8AVABTAEUAXwA5ADcALgBwAGQAZgAPABoADABNAGEAYwBpAG4AdABvAHMAaAAgAEgARAASADBVc2Vycy9qYWNreS9Eb3dubG9hZHMvUGFwZXJzL1NoZXBwZXJkX1RTRV85Ny5wZGYAEwABLwAAFQACAAz//wAAgAXSHB0eH1gkY2xhc3Nlc1okY2xhc3NuYW1lox8gIV1OU011dGFibGVEYXRhVk5TRGF0YVhOU09iamVjdF8QNi4uLy4uL1VzZXJzL2phY2t5L0Rvd25sb2Fkcy9QYXBlcnMvU2hlcHBlcmRfVFNFXzk3LnBkZtIcHSQloiUhXE5TRGljdGlvbmFyeRIAAYagXxAPTlNLZXllZEFyY2hpdmVyAAgAEQAWAB8AKAAyADUAOgA8AEUASwBSAF0AZQBsAG8AcQBzAHYAeAB6AHwAhgCTAJgAoAJAAkICRwJQAlsCXwJtAnQCfQK2ArsCvgLLAtAAAAAAAAACAQAAAAAAAAAoAAAAAAAAAAAAAAAAAAAC4g==},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/32.637387}}

@inproceedings{khaled06,
	Abstract = {Software project cost and effort estimation has become an increasingly important field in the past years due to the overwhelming role of software in today's global market. Several studies have been dedicated to create models in order to estimate the effort of software development. Most of the studies focused on expert judgment, analogy, parametric and algorithmic methods, bottom-up methods, and top-down methods. Nearly all estimating methods need information about how projects have been implemented in the past. However, this information may be of limited use to estimators, as there are uncertainties in the way that various terms, variables and factors are being interpreted. Two projects that may seem similar may indeed be different in a critical way. Moreover, the uncertainty in assessing similarities and differences means that two different analysts could develop significantly different views and effort estimates. The major contributions this paper makes are: 1) identification of an ontology-based cost estimation process framework for defining the semantics of project development data; 2) introduce the culture factor as it affects the software effort estimation; and 3) development of a software effort estimation ontology system (SEEOS) for use in estimating software project cost in a group of organizations. The system establishes a set of common project parameters between different projects and provides a common understanding of project parameters and their semantics. This system enables project managers to elicit software project features that are semantically compatible with new project requirements. The system has been implemented using Java and a relational database management system and data which have been collected from within UAE companies using an online system},
	Author = {Khaled Hamdan and Hazem El Khatib and John Moses and Peter Smith},
	Booktitle = {Innovations in Information Technology, 2006},
	Date-Modified = {2010-08-03 16:36:52 +1000},
	Doi = {10.1109/INNOVATIONS.2006.301942},
	Keywords = {Java;case-based reasoning;online system;project development data semantics;project requirement;relational database management system;software cost ontology system;software development;software project cost estimation;software project effort estimation;software project feature elicitation;Internet;Java;case-based reasoning;ontologies (artificial intelligence);project management;relational databases;software cost estimation;software development management;},
	Month = {nov.},
	Pages = {1 -5},
	Title = {A Software Cost Ontology System for Assisting Estimation of Software Project Effort for Use with Case-Based Reasoning},
	Year = {2006},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/INNOVATIONS.2006.301942}}

@inproceedings{idri02,
	Abstract = { Estimation models in software engineering are used to predict some important attributes of future entities such as software development effort, software reliability and programmers' productivity. Among these models, those estimating software effort have motivated considerable research in recent years. The prediction procedure used by these software-effort models can be based on a mathematical function or other techniques such as analogy based reasoning, neural networks, regression trees, and rule induction models. Estimation by analogy is one of the most attractive techniques in the software effort estimation field. However, the procedure used in estimation by analogy is not yet able to handle correctly linguistic values (categorical data) such as 'very low', 'low' and 'high'. We propose a new approach based on reasoning by analogy, fuzzy logic and linguistic quantifiers to estimate software project effort when it is described either by numerical or linguistic values; this approach is referred to as Fuzzy Analogy. This paper also presents an empirical validation of our approach based on the COCOMO'81 dataset.},
	Author = {Idri, A. and Abran, A. and Khoshgoftaar, T.M.},
	Booktitle = {Software Metrics, 2002. Proceedings. Eighth IEEE Symposium on},
	Date-Modified = {2010-08-03 16:35:51 +1000},
	Doi = {10.1109/METRIC.2002.1011322},
	Keywords = {COCOMO dataset; Fuzzy Analogy; analogy based reasoning; estimation by analogy; fuzzy logic; linguistic quantifiers; linguistic values; programmer productivity; reasoning by analogy; software engineering estimation models; software metrics; software project effort estimation; software reliability; case-based reasoning; fuzzy logic; project management; software development management; software metrics;},
	Pages = {21 - 30},
	Title = {Estimating software project effort by analogy based on linguistic values},
	Year = {2002},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/METRIC.2002.1011322}}

@inproceedings{hihn91,
	Abstract = {The authors describe a survey conducted at the Jet Propulsion Laboratory (JPL) to estimate software costs for software intensive projects in JPL's technical divisions. Respondents to the survey described what techniques they use in estimating software costs and, in an experiment, each respondent estimated the size and cost of a specific piece of software described in a design document provided by the authors. It was found that the majority of the technical staff estimating software costs use informal analogy and high-level partitioning of requirements, and that no formal procedure exists for incorporating risk and uncertainty. The technical staff is significantly better at estimating effort than size. However, in both cases the variances are so large that there is a 30% probability that any one estimate can be more than 50% off},
	Author = {Hihn, J. and Habib-agahi, H.},
	Booktitle = {Software Engineering, 1991. Proceedings., 13th International Conference on},
	Date-Modified = {2010-08-03 16:34:23 +1000},
	Doi = {10.1109/ICSE.1991.130653},
	Keywords = {cost estimation;design document;high-level partitioning;software intensive projects;survey;economics;software engineering;},
	Month = {13-16},
	Pages = {276 -287},
	Title = {Cost estimation of software intensive projects: a survey of current practices},
	Year = {1991},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ICSE.1991.130653}}

@inproceedings{keung08a,
	Abstract = {We developed a novel method called Analogy-X to provide statistical inference procedures for analogy- based software effort estimation. Analogy-X is a method to statistically evaluate the relationship between useful project features and target features such as effort to be estimated, which ensures the dataset used is relevant to the prediction problem, and project features are selected based on their statistical contribution to the target variables. We hypothesize that this method can be (1) easily applied to a much larger dataset, and (2) also it can be used for incorporating joint effort and duration estimation into analogy, which was not previously possible with conventional analogy estimation. To test these two hypotheses, we conducted two experiments using different datasets. Our results show that Analogy-X is able to deal with ultra large datasets effectively and provides useful statistics to assess the quality of the dataset. In addition, our results show that feature selection for duration estimation differs from feature selection for joint-effort duration estimation. We conclude Analogy-X allows users to assess the best procedure for estimating duration given their specific requirements and dataset.},
	Author = {Keung, J. and Kitchenham, B.},
	Booktitle = {Software Engineering, 2008. ASWEC 2008. 19th Australian Conference on},
	Date-Modified = {2010-08-03 16:37:24 +1000},
	Doi = {10.1109/ASWEC.2008.4483211},
	Issn = {1530-0803},
	Keywords = {analogy-X;datasets;duration estimation;feature selection;joint-effort estimation;software cost estimation;statistical inference;software cost estimation;statistical analysis;},
	Month = {26-28},
	Pages = {229 -238},
	Title = {Experiments with Analogy-X for Software Cost Estimation},
	Year = {2008},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ASWEC.2008.4483211}}

@inproceedings{shepperd96,
	Abstract = {The staff resources or effort required for a software project are notoriously difficult to estimate in advance. To date most work has focused upon algorithmic cost models such as COCOMO and Function Points. These can suffer from the disadvantage of the need to calibrate the model to each individual measurement environment coupled with very variable accuracy levels even after calibration. An alternative approach is to use analogy for estimation. We demonstrate that this method has considerable promise in that we show it to out perform traditional algorithmic methods for six different datasets. A disadvantage of estimation by analogy is that it requires a considerable amount of computation. The paper describes an automated environment known as ANGEL that supports the collection, storage and identification of the most analogous projects in order to estimate the effort for a new project. ANGEL is based upon the minimisation of Euclidean distance in n-dimensional space. The software is flexible and can deal with differing datasets both in terms of the number of observations (projects) and in the variables collected. Our analogy approach is evaluated with six distinct datasets drawn from a range of different environments and is found to outperform other methods. It is widely accepted that effective software effort estimation demands more than one technique. We have shown that estimating by analogy is a candidate technique and that with the aid of an automated environment is an eminently practical technique},
	Author = {Shepperd, M. and Schofield, C. and Kitchenham, B.},
	Booktitle = {Software Engineering, 1996., Proceedings of the 18th International Conference on},
	Date-Modified = {2010-08-03 16:34:35 +1000},
	Doi = {10.1109/ICSE.1996.493413},
	Keywords = {ANGEL automated environment;Euclidean distance minimisation;analogous project collection;analogous project identification;analogous project storage;analogy;computation;datasets;effort estimation;n-dimensional space;observations;software project;variables;project management;software cost estimation;software development management;software tools;},
	Month = {25-29},
	Pages = {170 -178},
	Title = {Effort estimation using analogy},
	Year = {1996},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ICSE.1996.493413}}
@article{Li2006,
	Author = {Li, J. and Ruhe, G.},
	Journal = {Proceedings of the 2006 ACM/IEEE international symposium on Empirical software engineering},
	Pages = {74},
	Publisher = {ACM},
	Title = {A comparative study of attribute weighting heuristics for effort estimation by analogy},
	Url = {http://portal.acm.org/citation.cfm?id=1159733.1159746},
	Year = {2006},
	Bdsk-Url-1 = {http://portal.acm.org/citation.cfm?id=1159733.1159746}}

@inproceedings{mendes00,
	Abstract = {Although estimating the effort required in developing Web applications is a difficult task, accurate estimates of development effort have an important role to play in the successful management of Web development projects. In software development work to date, emphasis has focused on algorithmic cost models such as COCOMO and function points. Two disadvantages of these models are firstly, the need for calibration of a model for each individual measurement environment and, secondly, the variable accuracy levels achieved even after calibration. The paper describes the use of estimation by analogy to calculate the development effort of Web applications. Two datasets containing empirical Web development data were used in the case study. One set contained data relating to forty-one novice developers, the other to twenty-nine experienced developers. The ANGEL tool supporting the automatic collection, storage and identification of the most analogous projects was used as a basis for estimating effort required for a new project. Results show estimation by analogy to be a promising alternative to algorithmic techniques},
	Author = {Mendes, E. and Counsell, S.},
	Booktitle = {Software Engineering Conference, 2000. Proceedings. 2000 Australian},
	Date-Modified = {2010-08-03 16:35:06 +1000},
	Doi = {10.1109/ASWEC.2000.844577},
	Keywords = {ANGEL tool;Web applications;Web development effort estimation;Web development projects;algorithmic cost models;analogous projects;analogy based estimation;automatic collection;calibration;datasets;development effort;empirical Web development data;measurement environment;novice developers;software development work;variable accuracy levels;distributed programming;information resources;project management;software cost estimation;software development management;},
	Pages = {203 -212},
	Title = {Web development effort estimation using analogy},
	Year = {2000},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ASWEC.2000.844577}}

@article{myrtveit05,
	Abstract = { Empirical studies on software prediction models do not converge with respect to the question "which prediction model is best?" The reason for this lack of convergence is poorly understood. In this simulation study, we have examined a frequently used research procedure comprising three main ingredients: a single data sample, an accuracy indicator, and cross validation. Typically, these empirical studies compare a machine learning model with a regression model. In our study, we use simulation and compare a machine learning and a regression model. The results suggest that it is the research procedure itself that is unreliable. This lack of reliability may strongly contribute to the lack of convergence. Our findings thus cast some doubt on the conclusions of any study of competing software prediction models that used this research procedure as a basis of model comparison. Thus, we need to develop more reliable research procedures before we can have confidence in the conclusions of comparative studies of software prediction models.},
	Author = {Myrtveit, I. and Stensrud, E. and Shepperd, M.},
	Date-Modified = {2010-08-03 16:36:35 +1000},
	Doi = {10.1109/TSE.2005.58},
	Issn = {0098-5589},
	Journal = {Software Engineering, IEEE Transactions on},
	Keywords = {accuracy indicator; analogy estimation; arbitrary function approximators; convergence; cost estimation; cross validation; data sample; empirical method; machine learning model; regression model; simulation; software metrics; software prediction model; software reliability; software validity; convergence; function approximation; learning (artificial intelligence); program verification; regression analysis; software cost estimation; software metrics; software reliability;},
	Month = {may},
	Number = {5},
	Pages = {380 - 391},
	Title = {Reliability and validity in comparative studies of software prediction models},
	Volume = {31},
	Year = {2005},
	Bdsk-File-1 = {YnBsaXN0MDDUAQIDBAUIJidUJHRvcFgkb2JqZWN0c1gkdmVyc2lvblkkYXJjaGl2ZXLRBgdUcm9vdIABqAkKFRYXGyIjVSRudWxs0wsMDQ4RFFpOUy5vYmplY3RzV05TLmtleXNWJGNsYXNzog8QgASABqISE4ACgAOAB1lhbGlhc0RhdGFccmVsYXRpdmVQYXRo0hgNGRpXTlMuZGF0YU8RAaQAAAAAAaQAAgAADE1hY2ludG9zaCBIRAAAAAAAAAAAAAAAAAAAAMgHQrtIKwAAABIP7BVNeXJ0dmVpdF9UU0VfMjAwNS5wZGYAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEhBiyHxZhgAAAAAAAAAAAAIABQAACSAAAAAAAAAAAAAAAAAAAAAGUGFwZXJzABAACAAAyAa2GwAAABEACAAAyHvM5gAAAAEAEAASD+wACEnkAAhJ1AAAkj8AAgA/TWFjaW50b3NoIEhEOlVzZXJzOmphY2t5OkRvd25sb2FkczpQYXBlcnM6TXlydHZlaXRfVFNFXzIwMDUucGRmAAAOACwAFQBNAHkAcgB0AHYAZQBpAHQAXwBUAFMARQBfADIAMAAwADUALgBwAGQAZgAPABoADABNAGEAYwBpAG4AdABvAHMAaAAgAEgARAASADJVc2Vycy9qYWNreS9Eb3dubG9hZHMvUGFwZXJzL015cnR2ZWl0X1RTRV8yMDA1LnBkZgATAAEvAAAVAAIADP//AACABdIcHR4fWCRjbGFzc2VzWiRjbGFzc25hbWWjHyAhXU5TTXV0YWJsZURhdGFWTlNEYXRhWE5TT2JqZWN0XxA4Li4vLi4vVXNlcnMvamFja3kvRG93bmxvYWRzL1BhcGVycy9NeXJ0dmVpdF9UU0VfMjAwNS5wZGbSHB0kJaIlIVxOU0RpY3Rpb25hcnkSAAGGoF8QD05TS2V5ZWRBcmNoaXZlcgAIABEAFgAfACgAMgA1ADoAPABFAEsAUgBdAGUAbABvAHEAcwB2AHgAegB8AIYAkwCYAKACSAJKAk8CWAJjAmcCdQJ8AoUCwALFAsgC1QLaAAAAAAAAAgEAAAAAAAAAKAAAAAAAAAAAAAAAAAAAAuw=},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/TSE.2005.58}}

@inproceedings{briand00,
	Abstract = {Delivering a software product on time, within budget, and to an agreed level of quality is a critical concern for many software organizations. Underestimating software costs can have detrimental effects on the quality of the delivered software and thus on a company's business reputation and competitiveness. On the other hand, overestimation of software cost can result in missed opportunities to funds in other projects. In response to industry demand, a myriad of estimation techniques has been proposed during the last three decades. In order to assess the suitability of a technique from a diverse selection, its performance and relative merits must be compared. The current study replicates a comprehensive comparison of common estimation techniques within different organizational contexts, using data from the European Space Agency. Our study is motivated by the challenge to assess the feasibility of using multi-organization data to build cost models and the benefits gained from company-specific data collection. Using the European Space Agency data set, we investigated a yet unexplored application domain, including military and space projects. The results showed that traditional techniques, namely, ordinary least-squares regression and analysis of variance outperformed analogy-based estimation and regression trees. Consistent with the results of the replicated study no significant difference was found in accuracy between estimates derived from company-specific data and estimates derived from multi-organizational data},
	Author = {Briand, L.C. and Langley, T. and Wieczorek, I.},
	Booktitle = {Software Engineering, 2000. Proceedings of the 2000 International Conference on},
	Date-Modified = {2010-08-03 16:35:12 +1000},
	Doi = {10.1109/ICSE.2000.870428},
	Keywords = {common software cost modeling;cost models;least-squares regression;replicated assessment;software costs;software organizations;software product;software cost estimation;},
	Pages = {377 -386},
	Title = {A replicated assessment and comparison of common software cost modeling techniques},
	Year = {2000},
	Bdsk-File-1 = {YnBsaXN0MDDUAQIDBAUIJidUJHRvcFgkb2JqZWN0c1gkdmVyc2lvblkkYXJjaGl2ZXLRBgdUcm9vdIABqAkKFRYXGyIjVSRudWxs0wsMDQ4RFFpOUy5vYmplY3RzV05TLmtleXNWJGNsYXNzog8QgASABqISE4ACgAOAB1lhbGlhc0RhdGFccmVsYXRpdmVQYXRo0hgNGRpXTlMuZGF0YU8RAaAAAAAAAaAAAgAADE1hY2ludG9zaCBIRAAAAAAAAAAAAAAAAAAAAMgHQrtIKwAAABIP7BRCcmlhbmRfSUNTRV8yMDAwLnBkZgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEhC1yHxaRwAAAAAAAAAAAAIABQAACSAAAAAAAAAAAAAAAAAAAAAGUGFwZXJzABAACAAAyAa2GwAAABEACAAAyHvNpwAAAAEAEAASD+wACEnkAAhJ1AAAkj8AAgA+TWFjaW50b3NoIEhEOlVzZXJzOmphY2t5OkRvd25sb2FkczpQYXBlcnM6QnJpYW5kX0lDU0VfMjAwMC5wZGYADgAqABQAQgByAGkAYQBuAGQAXwBJAEMAUwBFAF8AMgAwADAAMAAuAHAAZABmAA8AGgAMAE0AYQBjAGkAbgB0AG8AcwBoACAASABEABIAMVVzZXJzL2phY2t5L0Rvd25sb2Fkcy9QYXBlcnMvQnJpYW5kX0lDU0VfMjAwMC5wZGYAABMAAS8AABUAAgAM//8AAIAF0hwdHh9YJGNsYXNzZXNaJGNsYXNzbmFtZaMfICFdTlNNdXRhYmxlRGF0YVZOU0RhdGFYTlNPYmplY3RfEDcuLi8uLi9Vc2Vycy9qYWNreS9Eb3dubG9hZHMvUGFwZXJzL0JyaWFuZF9JQ1NFXzIwMDAucGRm0hwdJCWiJSFcTlNEaWN0aW9uYXJ5EgABhqBfEA9OU0tleWVkQXJjaGl2ZXIACAARABYAHwAoADIANQA6ADwARQBLAFIAXQBlAGwAbwBxAHMAdgB4AHoAfACGAJMAmACgAkQCRgJLAlQCXwJjAnECeAKBArsCwALDAtAC1QAAAAAAAAIBAAAAAAAAACgAAAAAAAAAAAAAAAAAAALn},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ICSE.2000.870428}}

@article{Li2008,
	Author = {Li, J and Ruhe, G},
	File = {:Users/ekremkocaguneli/Library/Application Support/Mendeley Desktop/Downloaded/Li, Ruhe - 2008 - Analysis of attribute weighting heuristics for analogy-based software effort estimation.pdf:pdf},
	Journal = {Empirical Software Engineering},
	Keywords = {attribute weighting,effort estimation by analogy,feature selection,heuristics,learning,rough set analysis},
	Number = {1},
	Pages = {63--96},
	Publisher = {Springer},
	Title = {{Analysis of attribute weighting heuristics for analogy-based software effort estimation method AQUA+}},
	Url = {http://www.springerlink.com/index/P821335293V53481.pdf},
	Volume = {13},
	Year = {2008},
	Bdsk-Url-1 = {http://www.springerlink.com/index/P821335293V53481.pdf}}

@article{Sentas2005,
	Author = {Panagiotis Sentas and Lefteris Angelis and Ioannis Stamelos and George Bleris},
	Doi = {DOI: 10.1016/j.infsof.2004.05.001},
	Issn = {0950-5849},
	Journal = {Information and Software Technology},
	Keywords = {Ordinal regression},
	Number = {1},
	Pages = {17 - 29},
	Title = {Software productivity and effort prediction with ordinal regression},
	Url = {http://www.sciencedirect.com/science/article/B6V0B-4CVR4MP-1/2/72e600ea322bfe4d64b1c90c3a968013},
	Volume = {47},
	Year = {2005},
	Bdsk-Url-1 = {http://www.sciencedirect.com/science/article/B6V0B-4CVR4MP-1/2/72e600ea322bfe4d64b1c90c3a968013},
	Bdsk-Url-2 = {http://dx.doi.org/10.1016/j.infsof.2004.05.001}}

@inproceedings{Lum2008b,
	Annote = {While we have found that COCOMO is a very robust model, our results also indicate that local calibration using boot strapping over standard regression, combined with variable reduction (column pruning) andstratification (row pruning using nearest neighbor) is in the vast majority of experiments the most efficient and effective tuning method.},
	Author = {Lum, Karen and Menzies, Tim and Baker, Dan},
	Date-Modified = {2010-08-20 15:40:33 +1000},
	File = {:Users/ekremkocaguneli/Library/Application Support/Mendeley Desktop/Downloaded/Lum, Menzies, Baker - 2008 - 2CEE, A TWENTY FIRST CENTURY EFFORT ESTIMATION METHODOLOGY.pdf:pdf},
	Journal = {International Society of Parametric Analysis (ISPA / SCEA)},
	Number = {May},
	Pages = {12 -- 14},
	Title = {2CEE, A Twenty First Century Effort Estimation Methodology},
	Year = {2008}}

@inproceedings{Kocaguneli2010,
	Address = {New York, NY, USA},
	Author = {Ekrem Kocaguneli and Gregory Gay and Ye Yang and Tim Menzies and Jacky Keung},
	Booktitle = {ASE '10: To Appear In the Proceedings of the twenty-second IEEE/ACM international conference on Automated software engineering},
	Title = {When to Use Data from Other Projects for Effort Estimation},
	Year = {2010}}

@inproceedings{Kultur2008,
	Address = {New York, NY, USA},
	Author = {Kultur, Yigit and Turhan, Burak and Bener, Ayse Basar},
	Booktitle = {SIGSOFT '08/FSE-16: Proceedings of the 16th ACM SIGSOFT International Symposium on Foundations of software engineering},
	File = {:Users/ekremkocaguneli/Library/Application Support/Mendeley Desktop/Downloaded/Kultur, Turhan, Bener - 2008 - ENNA software effort estimation using ensemble of neural networks with associative memory.pdf:pdf},
	Isbn = {978-1-59593-995-1},
	Pages = {330--338},
	Title = {{ENNA: software effort estimation using ensemble of neural networks with associative memory}},
	Year = {2008}}

@inproceedings{Turhan2007,
	Author = {Turhan, B. and Kutlubay, O. and Bener, A.},
	Booktitle = {Empirical Software Engineering and Measurement, 2007. ESEM 2007. First International Symposium on},
	Doi = {10.1109/ESEM.2007.57},
	Issn = {1938-6451},
	Keywords = {Isomap;feature extraction method evaluation;nonlinear feature extraction;principal component analysis;software cost estimation performance;support vector regression;feature extraction;principal component analysis;regression analysis;software cost estimation;},
	Month = {20-21},
	Pages = {497 -497},
	Title = {Evaluation of Feature Extraction Methods on Software Cost Estimation},
	Year = {2007},
	Bdsk-Url-1 = {http://dx.doi.org/10.1109/ESEM.2007.57}}

@article{Finnie1997,
	Author = {G. R. Finnie and G. E. Wittig and J-M. Desharnais},
	Doi = {DOI: 10.1016/S0164-1212(97)00055-1},
	Issn = {0164-1212},
	Journal = {Journal of Systems and Software},
	Number = {3},
	Pages = {281 - 289},
	Title = {A comparison of software effort estimation techniques: Using function points with neural networks, case-based reasoning and regression models},
	Url = {http://www.sciencedirect.com/science/article/B6V0N-3SP2RBC-6/2/4416e483f8e9fa78f01486d8fa6b7693},
	Volume = {39},
	Year = {1997},
	Bdsk-Url-1 = {http://www.sciencedirect.com/science/article/B6V0N-3SP2RBC-6/2/4416e483f8e9fa78f01486d8fa6b7693},
	Bdsk-Url-2 = {http://dx.doi.org/10.1016/S0164-1212(97)00055-1}}
