CHips L MINI SHELL

CHips L pro

Current Path : /proc/2/root/proc/2/cwd/usr/local/ssl/share/doc/cvs-1.11.23/
Upload File :
Current File : //proc/2/root/proc/2/cwd/usr/local/ssl/share/doc/cvs-1.11.23/PROJECTS

This is a list of projects for CVS.  In general, unlike the things in
the TODO file, these need more analysis to determine if and how
worthwhile each task is.

I haven't gone through TODO, but it's likely that it has entries that
are actually more appropriate for this list.

0. Improved Efficency

* CVS uses a single doubly linked list/hash table data structure for
  all of its lists.  Since the back links are only used for deleting
  list nodes it might be beneficial to use singly linked lists or a
  tree structure.  Most likely, a single list implementation will not
  be appropriate for all uses.

  One easy change would be to remove the "type" field out of the list
  and node structures.  I have found it to be of very little use when
  debugging, and each instance eats up a word of memory.  This can add
  up and be a problem on memory-starved machines.

  Profiles have shown that on fast machines like the Alpha, fsortcmp()
  is one of the hot spots.

* Dynamically allocated character strings are created, copied, and
  destroyed throughout CVS.  The overhead of malloc()/strcpy()/free()
  needs to be measured.  If significant, it could be minimized by using a
  reference counted string "class".

* File modification time is stored as a character string.  It might be
  worthwile to use a time_t internally if the time to convert a time_t
  (from struct stat) to a string is greater that the time to convert a
  ctime style string (from the entries file) to a time_t.  time_t is
  an machine-dependant type (although it's pretty standard on UN*X
  systems), so we would have to have different conversion routines.
  Profiles show that both operations are called about the same number
  of times.

* stat() is one of the largest performance bottlenecks on systems
  without the 4.4BSD filesystem.  By spliting information out of
  the filesystem (perhaps the "rename database") we should be 
  able to improve performance.

* Parsing RCS files is very expensive.  This might be unnecessary if
  RCS files are only used as containers for revisions, and tag,
  revision, and date information was available in easy to read 
  (and modify) indexes.  This becomes very apparent with files
  with several hundred revisions.

1. Improved testsuite/sanity check script

* Need to use a code coverage tool to determine how much the sanity
  script tests, and fill in the holes.


Copyright 2K16 - 2K18 Indonesian Hacker Rulez