I am looking at trying to use ferret/aaf to supplement my querying
against a
medium and large table with lots of columns. Some facts first:
Ferret 0.11.4
AAF 0.4.0
Ruby 1.8.6
Rails 1.2.3
Medium table:
105,464 rows
168 columns (mostly varchar(20))
11 actual columns indexed in aaf plus
40 virtual columns indexed in aaf (virtual is concat of two physical
columns.
e.g. cast_first_name_1 + cast_last_name_1 through cast_first_name_20 +
cast_last_name_20)
Large table:
1,244,716 rows
same column/index structure
These tables are not updated via Ruby, only read. I am trying to use
rebuild_index to bootstrap the medium sized table and it is taking a
very long
time (running for about 4 hours, indicates 50% complete with 4 hours
remaining)
and creating a massive number of files in the index directory (currently
about
65k, was 90k earlier)
I have not done any tuning of ferret/aaf so far, and I fear what it will
look
like to do the big table. Does anyone have any advise on how to speed
this
process up? Because the tables are updated by an external batch
process, if I
were to continue down this ferret/aaf path, I’d have to be looking at
running
this rebuild_index a couple of times per week which would be rather
painful
given the present time and might not be possible if the large table took
more
than 48 hours…