Course Home | Course Policies | Homework | Lab Open Hours | Programming | Labs | Schedule & Lecture Notes

CS 180: Data Structures
Spring 2011
MTuWF 10-10:50am in 121 RH

Erin Chambers
Contact Info: echambe5 - at -
Office: 011 Ritter Hall
Office Hours: TBA

The table below lists the lab assignments and associated dates.

Lab Topic Class Date Due Collaboration Policy
Lab 1: Copier Reduction C++ Tuesday, January 25 Friday, January 28 by 11:59pm pair (as needed)
Lab 2: Speed Limit C++: Loops and state information Tuesday, February 1 Friday, February 4 by 11:59pm pair (as needed)
Lab 3: Symmetric Order Stacks Wednesday, February 9 Friday, February 11 by 11:59pm pair (as needed)
Lab 4: Doubles Arrays Tuesday, February 15 Friday, February 18 by 11:59pm pair (as needed)
Lab 5: Symmetric Order revisited Vectors Tuesday, February 22 Friday, February 25 by 11:59pm pair (as needed)
Lab 6: Tanning Salon Arrays Tuesday, March 8 Friday, March 11 by 11:59pm pair (as needed)
Lab 7: Booklet Recursion Tuesday, March 22 Friday, March 25 by 11:59pm pair (as needed)
Lab 8: Trees Recursion and Trees Tuesday, March 29 Wednesday, April 6 by 11:59pm pair (as needed)

Information About Lab Assignments

ACM's International Collegiate Programming Contest

This year, each of our lab assignments will be a problem taken from a past offering of the ACM International Collegiate Programming Contest (ICPC). Students work in teams of three to solve as many problems as possible in a five-hour time period.

Hundreds of regional contests are held each Fall, involving thousands of teams across the world. The top 100 teams from the regionals qualify for the World Finals held in the Spring. More information can be found on the official ICPC site. Also, if you have any interest in participating on SLU's teams, please let me know next Fall.

General Problem Format

Each problem is computational in nature, with the goal being to compute a specific output based on some input parameters. Each problem defines a clear and unambigous form for the expected input and desired output. Relevant bounds on the size of the input are clearly specified. To be successful, the program must complete within 60 seconds on the given machine (thus, efficiency can be important for certain problems).

Each problem description offers a handful of sample inputs and the expected output for those trials as a demonstration. Behind the scene, the judges often have hundreds of additional tests. Submitted programs are "graded" by literally running them on all of the judges' tests, capturing the output, and comparing whether the output is identical (character-for-character) to the expected output.

If the test is successful, the team gets credit for completing the problem. If the test fails, the team is informed of the failure and allowed to resubmit (with a slight penalty applied). However, the team receives very little feedback from the judges. In essence, they are told that it failed but given no explanation as to the cause of the problem, or even the data set that leads to the problem.

Actually, the feedback is slightly more informative. Upon submitting a program, the team formally receives one of the following responses:

Important Conventions

Because of the automated nature of the judging, it is important that programs follow these conventions:

Please note as well that the format of most problems is designed so that judges can specify multiple tests as part of a single execution. Typically this is done by having an input format where initial parameters are read, using a special value (such as 0 or #) to designate the end of the trials. Therefore, most programs will need to iterate through multiple trials using an outer loop. It is also important that relevant data structures be initialized for each trial, so that earlier trials do not affect later ones.

Testing Your Implementation

Formal Submission

Although the online tools give you a way to test your implementation, you must submit to the instructor via email in order to get credit for the lab. Please keep in mind that we will award half-credit for any sincere attempt at a lab, even if you are unable to succeed with the automated testing. However, you will get zero credit if you do not ever submit your attempt.

Also, make sure to include the names of all team members in comments at the top of the submit source code.