- #1
Iron_Brute
- 19
- 0
I'm having a problem understanding what it is N stands for. The question is
"Write a program that will find the smallest, largest, and average values in a collection of N numbers. Get the value of N before scanning each value in the collection of N numbers."
To me it looks like N can stand for 2 different values, it could be the total number of values the user would input but it could also represent the individual numbers that the user inputs.
Im writing this in C#, and finding the smallest, largest, and average values I am fairly confident I can do that it is just what the N is supposed represent that is throwing me off.
"Write a program that will find the smallest, largest, and average values in a collection of N numbers. Get the value of N before scanning each value in the collection of N numbers."
To me it looks like N can stand for 2 different values, it could be the total number of values the user would input but it could also represent the individual numbers that the user inputs.
Im writing this in C#, and finding the smallest, largest, and average values I am fairly confident I can do that it is just what the N is supposed represent that is throwing me off.