User Forum
Xequte User Forums
Profile    Active Topics    Forum FAQ    Search this forumSearch
     
Forum Tips
» This forum is for users to discuss their software problems and suggestions for improvement.  Remember that only a tiny percentage of our users experience such problems (To be precise less than 0.7% have ever posted on this forum)
» Include your software version number (as displayed in the About Box)
» Check the FAQ before posting as most common questions have already been answered
» You don't need to also e-mail us separately as our support team will respond to all questions on this forum
 
 All Forums
 Xequte User Forums
 MailList King
 Members Belonging to Multiple Groups

Note: You must be registered in order to post a reply.

Size 
UserName:
Password:
Mode 
Format  BoldItalicizedUnderline Align LeftCenteredAlign Right Horizontal Rule Insert HyperlinkInsert EmailBrowse for an image to attach to your post Insert CodeInsert QuoteInsert List
   
Message 

 

Emoji
Smile [:)] Big Smile [:D] Cool [8D] Blush [:I]
Tongue [:P] Evil [):] Wink [;)] Black Eye [B)]
Frown [:(] Shocked [:0] Angry [:(!] Sleepy [|)]
Kisses [:X] Approve [^] Disapprove [V] Question [?]

 
Check here to subscribe to this topic.
   

T O P I C    R E V I E W
jimbo69 Posted - Apr 27 2010 : 10:38:48
Been trying to repair our database, and have exported and imported but don't understand members belonging to multiple groups. The master file only contains one group id. Not sure what file contains the other group membership information. 1.4million members

Still slow selecting groups even on a 3.6ghzx4 machine with raid. Does multiple group membership create additional indexes slowing things down more? I've tried repairing the old database but some corruption in the sent and task files rebuilding fails, might be a able to fix them with db utility, but have not tried that yet, takes a long time for each task.

I can try splitting the database, but configuration factors are involved. Over 100 groups and we still need to clean the db only about 30% are still valid so splitting and rejoining would be crazy.

If you have a database file definition specifications that would help just point me there.
Thanks
2   L A T E S T    R E P L I E S    (Newest First)
xequte Posted - Apr 28 2010 : 03:05:45
Hi

It sure sounds like you're having a difficult time.

Because it is an old version of MLK, I'm not able to test your import file to see why it might be failing. But an export file of that size can cause problems for Notepad and will definitely not load in Excel (which has a limit of 64,000 rows).

Can you confirm that you have limited loading of records into memory to 500 or less (under View > Options, Other)?

If so, what particular actions in MLK are slow?


Nigel
Xequte Software
www.xequte.com
nigel@xequte.com
jimbo69 Posted - Apr 28 2010 : 01:27:00
After beating it to death I finally was able to get the exported data in the new database. I exported a csv file from the old database, When I tried to import it into the new database it would die around 400,000 names. So I split the csv file funny that excel would not load the file. Notepad froze ending up using EditPad Lite a great little editor. You can turn line numbers on so it makes it easy to select say 0-300,000 names, cut them and paste them into a new files. I did this for our 1.4m database creating 5 files. Sometime MLK freezes when it encounters a orphaned member. If that happens I repair the main file with the database utilities and retry importing the file.

I don't know how the groups are handled but the members of multiple groups magically appear in each group even though the data being imported does not seem to contain multiple group data. I'm just glad I have this working.

I'm running the old version on window 7 so some of the errors might be due to that I do have it running in xp mode, definitely need to upgrade to the latest version waiting on the higher up's to approve the purchase.

I hope the latest version does better indexing when selecting groups and displaying the emails address's. But when I had only 400,000 entered it was much faster than 1.4m, Considering splitting the db