Messages - DK2IT [ switch to compact view ]

Pages: prev1 [2] 3next
6
Well, you can try to get something similar with tasklist and an AWK for windows or similar (like advanced calculation using FOR command).
For task manager, try Process Hacker, work quite well.

7
First time also I have got several error with PowerShell and this script  >:(
If I remember well, you must enable some authorization, and when you launch by commandline you must enter the FULL path of the script.

8
To get the total memory of a multi-process you can use a little PowerShell script (taken from here)
$chrome = get-process chrome -ErrorAction SilentlyContinue
if ($chrome -ne $null)
{
$m = ps chrome | measure PM -Sum ; ("chrome Physical Memory {0:N2}MB " -f ($m.sum / 1mb))
$m = ps chrome | measure WS -Sum ; ("chrome Working Set {0:N0}MB " -f ($m.sum / 1mb))
}
I've found that Working Set seem to give a better indication of real memory occupation.
Of course you can change process name and obtain memory usage for others multi-process applications, like opera, iexplore etc.
(with this system you can also sum several process like Firefox + plugin-container + FlashPlayerPlugin)

Or, you can use Process Hacker, the latest release (v2.36) include a tray popup window (can be sticky) that show the memory / cpu usage of most active processes,
and for multi-process app show the sum of all process (also for cpu!).

9
Where many of the entries are variations on the same base, user01 user02 user1979 user1980 etc..  my last suggestion would be only store the "base" of the dictionary entry and generate the variations.
And that can be an interesting idea  :Thmbsup:

10
Of course this is a quick and fast solution for bhuiraj, it's not optimal, but do not require special software. I've tested 1.5Gb of data for over 227 millions of words, and the DB is quite big and the search are not so fast. But, of course, if you need speed you can use DB like mysql or oracle using a fine tuned configuration (memory index, cache query, partition table, etc.).
In this case, however, is possible create an optimal solution (without the generic DB overhead), but you need to create a specific software to handle a very very big dictionary.

Pages: prev1 [2] 3next
Go to full version