Topics - peter.s [ switch to compact view ]

Pages: [1] 2 3 4 5next
1
First of all, and as you know, my working system is XP, and with "only" 2 giga of memory; I acknowledge that in/for 2015, this is decidedly sub-standard.

But then, as you know, too, there have been several threads here in this forum, and many more elsewhere, which treat the incredible sub-standardness of FireFox, re its inexistent memory management.

As said, I'm not into defamation, so I have to admit that in part, my probs could come from Avira free (formerly, I had used Avast free, even more intruding than Avast free), and also, I have to admit that my probs started with the very latest Adobe Flash update (16), which they offered in order to overcome (again, and, felt, for the 1,000th time) "security" probs.

I had installed that Flash 16, and then, after opening just SOME tabs in FF, I quickly not only ran out of memory, but had my system stalled for good, up to killing FF "process" by Win Task Manager, and by thus losing any tab = all the work I previously had put into searching for url's, links, etc. - it should be obvious for any reader that by opening some 12 or 15 tabs from search and links in previous "hits", you've got some "work done", which is quite awful to lose then.

I've always said, "you get what you pay for", and I've always acknowledged there are just SOME exception to that rule, but then, ALL of my experience backs this up, to 99.5 (= not: 99,9) p.c. of all cases, this rule applies perfectly, and FireFox seems to be the perfect example of TOTAL CRAP, delivered by some "volunteers", who like the idea that they are "giving out something valid for free", when in fact, they tell us, hey, dude, I know I cannot sell my shit, but ain't you willing to swallow it for free? Of course, I'm opening this thread not in order to defame FF, but in order to get new ideas about how to do things better, this whole forum being about that, right?

Thus, my very first reaction to FF being stalled* by that infamous Flash update was to deactivate Flash, and to observe things coming from that, for some week or so. Here I've got news for you: Flash, except for YT, is totally unnecessary, AND it's omnipresent (= "ubiquitous"), i.e. almost ANY web site, as poor-in-content or modest in scope it might be, there's virtually ALWAYS that line above my FF content window, "Do you allow FF to activate Flash for this site?" (or something like that, DC NOT doing this shit).

*= Of course, I've got plenty of room for "virtual memory M" by Windows, on c: (since my data, as said, is on some external hdd), and "virtual memory is managed by the system") - but notwithstanding, even if I allow a quarter of an hour (!!!) for any command to become effective, I always end up by killing the FF "process", after HOURS of waiting. At the same time, all other applications functions "quite normally", i.e. they respond to commands, but with that little delay you'd expect by my system's having replaced working memory by multiple hdd accesses, considering FF has eaten all the working memory. It's just FF that doesn't respond any at all.

And fact is, in more than a week, I NEVER had to tell FF to activate Flash, in order to get ANY useful info, from any of those several hundred pages all begging for Flash. (It's understood that for JavaScript, the situation is totally different: If you don't allow for JS, almost any web page of today will not work anymore, in any acceptable way. But again, don't mix up JS and Flash; JS having become a literally unavoidable "standard", whilst Flash is a simple nuisance, except for YT, and then, for rare cases in which you want to see some embedded "film" - IS propaganda? No thanks, and all the rest, no thank you either; let alone for indecently idiotic porn.)

Back to FF: My getting rid of Flash did NOT solve my probs. It's invariably "CPU 100 p.c." over hours, with Flash de-activated though, and as soon as I've got opened more than just 10 or 12 FF tabs; I assume these are JS scripts running, but then, even after MANY minutes, FF never tells me, "that JS is running, should we stop it?".

I have to say that I know about the existence of "NoScript for FF", but then, it's not obvious how to run that NS in some smooth way, just in order to intercept too-demanding scripts whenever they dare run, but leaving alone any menu "scripting" anywhere; do you

I wish to confirm again that I'm NOT speaking of porn or other crap sites, but that I'm just "surfing" among the most innocuous web sites you could imagine.

As for Flash, before deactivating Flash for good, I had tried Chrome, and I had the very unpleasant experience that with Chrome, and that incredible shit of Flash 16, all was as incredible awful as with FF and that incredible shit of Flash 16 (sic), if not worse (!), so it's obvious that Flash 16 is even worse than FF 36 (or was it 35? it's the current version all the same), but then, Chrome will allow your killing ONE tab running, whilst in FF, it's "all or nothing", i.e. if you decide to kill the FF "process", you will lose all your "search" work, too (since FF stalls your FF process (i.e. not your system as a whole, so it's obvious it's all a matter of FF's memory management), so it's not even possible to switch from one tab to another one in order to retrieve the respective url's, even manually).

Btw., WITH that incredible Flash 16, simple Flash sites (which in fact would not have even needed Flash to begin with, see above) brought FF to 1200 meg, then 1,500, then 2,000, 3,500 meg... in fact, Flash's memory demands are simply unlimited, and that's confirmed not currently (I admit), but from Flash users' experience back in August, 2014, i.e. some few Flash versions ago, and who say Flash of summer 2014 asked for unlimited memory, 6 giga, 8 gia, 10 giga... they were on systems of 8 or 16 giga of working memory, and they thought it was unbearable...

The only reason I cling to FF is the fact that "YouTube Video and Audio Downloader" is available for FF only (i.e. not Chrome), and that it's the ONLY YT downloader of my knowledge which lets you select best AUDIO quality, too (and not only best video quality, as its competitors do, at best) - but in the end, you can perfectly use FF for this YT downloading, whilst using Chrome for anything else, so that's "no reason".

Hence :

- Except for very limited usage (YT), Flash is totally useless and, short of viruses, the utmost nuisance on pc (or Mac) (and as usual, Jobs was first to identify this problem, AND to resolve it, for much of the systems he's been marketing)
- ( Similar things could be said about the really ridiculous and useless Adobe pdf viewer, but that's another story. )
- FF is to be considered liquid, stinking, green, morbid shit: If not even in iteration 36, software meets most basic standards, it will probably not meet them in iteration 100 either
- Chrome is "free", too, but we all know you pay with all your data... BUT: At least there, you KNOW WHAT price you pay for their "free" service, whilst FF "do it all benevolently", and obviously serve you perfect crap (whatever the reasons of FF being totally stuck, with 2 giga of work memory, and plenty of "virtual memory", your only alternative is to kill FF throughout if ever you want to get rid of some "CPU 100 p.c." over many, many minutes, with no end, instead of killing JUST SOME tabs going bonkers, is kindergarten)
- And yes, Avira free could be "in it" to some degree, too (= I had less problems, even with FF, when I "surfed" without any "protection") (but Avast free was really "unbearable", by their pop-ups (i.e. at least, I thought so, before my current problems with FF)... but perhaps, function-wise, they would always be preferable to Avira free, which is less intruding re pop-ups, but doesn't work as well with FF, then?)
- Any insight into NoScript for FF? Is there a chance to get it to stop JS scripts running amok but letting go of any "regular" JS script anywhere?

Your opinion/advice/experience is highly welcome.

EDIT:

Sorry, my mistake above, I just read:

"Allow www.donationcoder.com to run "Adobe Flash"?" - Should we not enter some overdue discussion re "Are site developers trying to do Flash even in pure-text pages utterly nuts?", right now?

2
Call me conservative; up to very recently I used two Nokia 9210i - why?

I

Two reasons, not at all related to each other, but equally important:

- I want a physical keyboard (ok, the Nokia kb is really bad, so this criterion is highly debatable), so the only other current alternatives would have been either other old smartphones (used ones), or that RIM stuff (changed their name but you know what I mean)

- I bought lots of expensive sw for those phones, and most readers will know that, it's smartphone sw developers who very early succeeded in forcing hardware linking (or what is it called?) to users: any mobile phone has got an IMEI number, and almost any (from my experience, 99 p.c. or more) sw for smartphones traditionally has been coupled to the IMEI in question: No (legal) chance even to de-install sw from phone 1 and THEN only install it to another phone: When your phone breaks, your expensive sw is dead.

I suppose this is also true for iPhones and Android (in fact I don't know), but the big difference is, there's a plethora of (also quite prof.) sw for both systems, costing between 2 and 15 bucks, when really useful smartphones-of-the-old-days sw came with prices much higher, and even into the 3 figures.

This being said, for sw developers, smartphones of the old days were a dream come true; it's just MS who today insist upon your sw licence being broken, together with your hardware, whilst decent sw-for-pc developers all allow for re-install when you change your hardware.

II

Now for batteries. As you will have guessed, I cannot use my (virtually "unbreakable": good old quality from the ancient times) Nokia phones anymore since I naïvely thought batteries would not become a problem, those "Communicators" having been sold by "millions", in very high numbers at the very least.

Well, I was wrong: Currently, they sell USED "Communicator" batteries for 3 figures, and my own little stock had come to an end, BEFORE I had figured out I should buy some additional supplies (and then, you cannot store "batteries" / cells (rechargebable or not) forever).

Ok, they now sell big batteries (and with quintupled capacity), with various adapters, even for those "Communicators", but buyer beware: Even if you're willing to use a smartphone connected with some crazy cable to some heavy battery in your pocket (well, in the old days a simple mobile phone was about 10 or 12 kg), this is not a solution since all (???) of these (= from their respective advertizing, not one will have the needed additional functionality indeed) will only work if you have got a healthy regular battery in your smartphone, too; in other words, the external battery can spice up your internal one, not replace it. Why do I know or think I now? (Perhaps I'm even mistaken???)

Now for the difference with many (all???) notebooks: I never had the slightest problem to connect my (over the years, multiple) notebooks to the current, and have them work fine, as long as the respective mains/power adapter was working correctly, long after the internal battery working and/or being available.

The same does not seem to be true with smartphones in general (???); at the very least, it's not true for my "Communicators":

It makes no difference if I have got a worn-out battery in the Nokia, or if I leave it out: Just connecting it to the power adapter (which in turn is connected to the mains of course, I'm not that a lunatic) will NOT do anything in order to my being able to start the phone, it remains just dead, and the same is true if I put the phone into its (equally expensive) "desk stand" (which in turn is connected to the power adapter). And since I've got two Nokias, several (worn-out) batteries, several power adapters, several desk stands, and know about permutations, I'm positive that my problems don't come from some broken smartphone.

In other words, my Nokias need a working internal battery in order to be able to take advantage from any external power supply, and from their respective ads, I suppose those external batteries will not make any difference; my question is, is this behavior typical for smartphones, or is it just typical for the dumbness of Nokia staff? (As we all know, Nokia is gone.)

If it's typical for mobile phones and / or smartphones in general, beware of investing too much into (even a well-sold) smartphone: Once you won't get any more batteries for that, all your investments in that phone will have been flushed.

III

So what I do for the time being? Went back to a combi of Nokia 6300 (har, har, batteries available as for now) and my old sub-notebook (with an internal umts card, reverting to "sleep state" in-between, and as long as the third-party cell will be alive) I hadn't really used any more for a long time:

Since those sub-notebooks are total crap: A regularly-sized notebook is difficult enough to type on (with 10 fingers, nor just 2 or 3) when in the office, you do right and use some decent, regular keybord, so it's obviously a very smart idea to buy some lightweight notebook for the road, but which has got a KB OF REGULAR SIZE (if not shape) - and don't forget the oh-so-useful (both for digit entering as for macroing!) dedicated keypad, and trust me about that; any sub-notebook (incl. those immensely pretty Sony sub-sub-notebooks that weren't continued though and now are available, used, for quadruple their price new) will be a constant and real pain-in-the-you-know-where: It's weight, not size that counts*, believe me, I'm judging from enough unpleasant first-hand experience.

IV

I just read, "Nikon kills third-party battery support", i.e. they probably put some additional electronics in their reflex camera preventing third-party battery makers from creating battery compatible cells: Another (for the consumer: very bad) "highly interesting" "development".


Your respective experiences / solutions would be very welcome.


*= this rule does not also apply in inter-human intimacy

3
...are not that superior either?

This is a spin-off from https://www.donationcoder.com/forum/index.php?topic=40074.0 discussing MR update from 5 to 6.

"I'm a big fan of macrium reflect.  Very fast, very stable, no bloat."

MR seems to be the premier backup-and-recovery sw on the market as far as the paid version is concerned (which is discussed above).

As for their free version, though, I only can encourage possible users to refrain from it, not because it was really bad (in fact, I never knew and don't know), but because it does not seem to offer any functionality going beyond what less-renowned competitors offer, in their respective free versions, or more precisely, it does offer even less than they do.

In fact, I went back to Paragon Backup and Recovery Free, where I can start to reinstall of my backup from within running Windows (which for that is than ended, then Linux will loaded for the rewrite of c: (or whatever), and then Windows is loaded again) - why should I fiddle around with doing lots of things manually, with MR (Free) if I can have this repeated os swapping, both by Paragon or EaseUS (and perhaps by others), done automatically?

MR (Free), on the other hand, did the backup (onto my hdd), and when I tried to reinstall that backup (after some bad experiences, I do such tries immediately after the original backup now, not weeks or months afterwards and hoping for the best in-between), it told me I didn't have an external reinstall device (or whatever they call it) from which to run the backup.

After this quite negative experience with MR (Free), I'm musing, of course, why MR (paid) is touted the way it is, since from the moment on you're willing to pay, you'll get incremental/differential backup/restore, from their competitors, too (Paragon, EaseUS and also Acronis: this latter I never touched, having read about very bad experiences from other users, allegedly having lost data with Acronic, and with several versions that is).

Also, MR did not seem anything "fast" to me, not faster than Paragon or EaseUS anyway, and at least for Paragon, I can say it's perfectly stable (I once lost data with their partition tool, but that was my fault, triggered by quite awful, quite ambiguous visuals in the respective Paragon program: So today I use Paragon for backup and EaseUS for partitioning).

And as an aside, MR even has got its own wikipedia entry, of which the wikipedia staff is far from being happy (and they say so), and which contains some direct links to the MR site where you would have expected links to less seller-specific info.

And to say it all, MR, on their homepage, currently advises you to update from 4 to 5, whilst above, it's said that 6 is imminent (?), and that updating from 5 to 6 is NOT free for v. 5 owners.

All this makes me think that perhaps MR do some very good pr and are able to create some hype, whilst at the end of the day, it's just a very regular, decent product which succeeded in realizing higher prices than their competitors are able to realize, by that hype.

If MR (paid) really has some usp(s), please name them; their free version at least is a lesser thing than their contenders' free products.

4
General Software Discussion / Scraper too expensive at 20 bucks
« on: January 16, 2015, 06:34 AM »
(Original post at bits and referred to here was "$19 + seems an awful lot of money for software you can get the same type of thing for nothing. (...)".)

The problem lies elsewhere. A price of 20 bucks is certainly not a deal breaker, neither would be 40 bucks (original price), and there are competitors that cost several hundred bucks, and which are not necessarily better or much better.

First,

if you search for "download manager", the web (and the people who constitute it by their respective contributions) mix up web scrapers (like A1) and tools for downloading files specified beforehand by the user, but the download of which will then be done within multiple threads, instead of just one, by this using your possible fast internet connection to its fullest; of course, most of the scrapers will include such accelerating functionality, too. Thus, the lacking discriminating effort in what commentators see as a "download manager" does not facilitate the discussion to begin with; you should perhaps use the terms "scrapers", and "download accelerators", for a start, but there is also some "middle thing", pseudo-scrapers who just download the current page, but without following its links.

Second,

the big problem for scrapers nowadays is Ajax and database techniques, i.e. many of today's web pages are not static anymore, but are built up from multiple elements coming from various sources, and you do not even see those scripts in full; scripts you can read by "see page source" refer back to scripts on their servers, and almost anything that is done behind these scenes, cannot be replicated by ANY scraper (i.e. not even by guessing parts of it, and from building up some alternative functionality from those guesses), so the remark that A1's pages from scraped Ajax pages do not "work" is meaningless.

The only other remark re A1 I found in the web was, you will get "the whole page", instead of just the photos, in case you would like to download just the photos of a web page; IF that is right, that was a weakness of A1 indeed, since these "choosing selected content only" questions are the core functionality today's scrapers could and should have, in the above-described general framework in which "original web page functionality" can not be replicated anymore, for many pages (which often are the ones which are of most interest = with the most money behind = with both the "best" content, and lots of money for ace programming).

Thus, "taking up" with server-side programming has become almost impossible for developers anyway, so they should revert to optimization of choosing selected content, and of making that content available, at least in a static way, and it goes without saying that multiple different degrees of optimization of that functionality are imaginable: built-in "macros" could replicate at least some standard connections between screen/data elements "on your side", and of which the original triggers are lost, by downloading, but this would involve lots of user-sided decisions to be made, and hence lots of dialogs the scraper would offer the user to begin with ("click on an element you want as a trigger, then select data (in a table e.g.) that would be made available from that trigger", or then, big data tables, which then you would hierarchically "sort" in groups, in order to make that data meaningful again).

It's clear as day that the better the guesses of the scraper in such scenarios, the easier such partial re-consitution of the original data would often become, and also, that programming such guesses-and-services-offered-from-those would both be very "expensive" in programming, and be a never-ending task, all this because today's web technologies succeed in hiding what's done on the server side.

In other words, from even very complicated but static, and even pseudo-dynamic (i.e. get it all out of databases, but in a stringent, easily-to-be-replicated way) web pages yesterday, to today's dynamic web pages, it has been a step beyond what scrapers sensibly would have been able to handle.

But it's obvious also that scrapers should at least perfectly handle "what they've got", and the above-mentioned example (as said, found in the web) of "just downloading the pics of a page", whilst being totally realistic, is far from being sufficient as a feature request:

In so many instances, the pics of the current page are either just thumbs, or then, just pics in some intermediate resolution, and the link to the full-resolution pic is not available but from the dedicated page of that middle-resolution pic, and the situation is further complicated by the fact that often, the first or second resolution is available, but the third resolution is not, and that within the same start page, i.e. for the same task at arrival, for some pics, the scraper / script would have to follow two or three links, in for other pics linked to at the same page, it would have to follow just one or two.

This being said, of course, such "get the best available resolution for the pics on current page" should be standard functionality for a scraper.

But, all this being said, it also appears as quite evident to me that for tasks beyond such "elaborate standard tasks" (and which could be made available by the scraper "guessing" possibly relevant links, then have the user choose from the intermediate results, and then the scraper building up the necessary "rule(s)" for the site in question), scraper programming comes with the additional problem that such "specific rule building" would be split into a) what the scraper would make available and b) what the user could make out of these pre-fetched instruments, whilst in fact, the better, easier, and ultimately far more powerful solution (because the limitations of the intermediate step would be done away, together with that intermediate step) would be to do scripting, but ideally having some library of standards at your disposal.

(Readers here in DC will remember my - unanswered - question here how to immediately get to "page x" (e.g. 50) of an "endless" Ajax page (of perhaps 300 such partial "pages" (or whatever you like to name those additions), instead of "endlessly" scrolling down to it.)

Anyway, precise selection of what the user wants to scrape, and of "what not", should be possible in detail, and not only for links to follow on start page, but also for links further down, at the very least for links "on page 2", i.e. on several kinds (!) of pages which only have in common the fact that all of them are one level "down" from the respective "start page" (I assume there are multiple but similar such "start pages", all of them to be treated in a similar (but not identical, see above) way.

Third,

so many scrapers (and download accelerators, too) tout their respective accelerating power, but few, if ever one, mention the biggest problem of them all: More and more server programs quickly throw your IP(s!) and even your PC out of their access scheme, should you dare scrape big content and/or, repeatedly, updated content, and again, as above, the more elaborate the content and their server-side page-build-up programming, the higher the chances are that they have sophisticated scraper detection, too.

What most people do not know, when they choose their tunnel provider, is the fact that in such "heavy-scraping" scenarios, it's quite "risky" to get a full-year contract (let alone something beyond a year), and that there are special tunnel providers where you rent multiple IPs at the same time instead - which comes at a price.

With these multiple addresses, many scraping guys think they are on the safe side - well, what's multiple addresses "abroad" (from the server's pov), and when in country x no such provider can provide you any, or more than just a handful of "national" IPs?

And it does not end there. How "visually good" is your script, from the server's pov again? Don't you think they cannot "put it all together again" when your scraping follows detectable rules? To begin with, your scraping is probably mutually exclusive, which is obviously a big mistake, but which facilitates combining the parts on your side, right? He, he...

And you're spacing your requests, of course, in order for the server not to detect it's a machine fetching the data? He, he, again, just spacing the requests in time does not mean the server will think it detects some real person, looking for the data in a way some bona fide prospect would look for that data.

Not to speak of bona fide prospects looking in certain standard ways, but which never are the same though, and that they don't do just sequential downloading ("sequential" does not mean, follow link 1, then 2, then 3, but link 35, 482, 47, whatever, but download, download, download!), but revert to some page before, press F5 here or there (but not systematically of course), and so on, and in endless ways: As soon as there is a possible script to be detected, those servers send a signal on a real person on their side, and who will then look into things, relying on their scripts-for-further-pattern-detection: time of the day for such a "session", amount of data downloaded, number of elements downloaded, order in which (sub-) elements are downloaded (patterns, too similar and/or or not "real-life" enough).

Then, even if you quite perfect all this, by having your machines replicating real-life behavior of different real persons, even most real-life prospects will not remain interested in the same or similar data over the years, and most of them, not even over months in a row!

And all this with the concurrent problem of the geographic repartition of your IPs again: Where almost all of their bona fide prospects would sit in some specific country, or even in some specific region of that country, and so all of the above problems, even if resolved in perfect ways (and this necessarily included lots of overlaps if you want your global scheme to remain "realistic") will be only partial solutions and not work for long if you cannot resolve the problem of how to fake IPs and their geography, instead of just renting some.

My 2 cent to put into perspective some naïve, "$19 + seems an awful lot of money for software you can get the same type of thing for nothing.", and I certainly left out additional aspects I didn't think of on the fly.

5
General Software Discussion / Desktop search; NTFS file numbers
« on: January 11, 2015, 07:55 AM »
This is a spin-off of page 32 (!) of this thread https://www.donationcoder.com/forum/index.php?topic=2434.775 ,

since I don't think real info should be buried within page 32 or 33 of a someday gross-page-long thread of which readers will perhaps read page 1, and then the very last (pages) only; on the other hand, even buried on some page 32, wrong and/or incomplete "info" should not be left unattended.
____________________

Re searching:

Read my posts in http://www.outlinersoftware.com/topics/viewt/5593

(re searching, and re tagging, the latter coming with the 260 chars for path plus filename limitations of course if you wanna do it within the file name... another possibly good reason to "encode" tags, in some form of .oac (Organisation(al things) - Assurances - Cars), instead of "writing them out")

Among other things, I say over there that you are probably well advised to use different tools for different search situations, according to the specific strengths of those tools; this is in accordance with what users say over here in the above DC thread.

Also, note that just searching within subsets of data is not only a very good idea for performance reasons (File Locator et al.), but also for getting (much) less irrelevant results: If you get 700 "hits", in many instances, it's not really a good idea to try to narrow down by adding further "AND" search terms, since that would probably exclude quite some relevant hits; narrowing down to specific directories would probably be the far better ("search in search") strategy; btw, another argument for tagging, especially for additional, specific tagging of everything that is in the subfolder into which it "naturally" belongs, but which belongs into alternative contexts, too (ultimately, a better file system should do this trick).

(Citations from the above page 32:)

Armando: "That said, I always find it weird when Everything is listed side by side with other software like X1, DTSearch or Archivarius. It's not the  same thing at all! Yes, most so called "Desktop search" software will be able to search file names (although not foldernames), but software like Everything won't be able to search file content." - Well said, I run into this irresponsible stew again and again; let's say that with "Everything" (and with Listary, which just integrates ET for this functionality), the file NAME search problem has definitely been resolved, but that does not resolve our full text search issues. Btw, I'm sure ET has been mentioned on pages 1 to 31 of that thread over and over again, and it's by nature such overlong threads will treat the same issues again and again, again and again giving the same "answers" to those identical problems, but of course, this will not stop posters who try to post just the maximum of post numbers, instead of trying to shut up whenever they can not add something new to the object of discussion. (I have said this before: Traditional forum sw is not the best solution for technical fora (or then, any forum), some tree-shaped sw (integrating a prominent subtree "new things", and other "favorites" sub-trees) would have been a thousand times better, and yes, such a system would obviously expose such overly-redundant, just-stealing-your-time posts. (At 40hz: Note I never said 100 p.c. of your posts are crap, I just say 95 or more p.c. of them are... well, sometimes they are quite funny at least, e.g. when a bachelor tries to tell fathers of 3 or 4 how to rise children: It's just that some people know-it-all, but really everything, for every thing in this life and this world, they are the ultimate expert - boys of 4 excel in this, too.)

Innuendo on Copernic: Stupid bugs, leaves out hits that should be there. I can confirm both observations, so I discarded this crap years before, and there is no sign things would have evolved in the right direction over there in the meantime, all to the contrary (v3>v4, OMG).

X1: See jity2's instructive link: http://forums.x1.com/viewtopic.php?f=68&t=9638) . My comment, though: X1's special option which then finds any (? did you try capitals, too, and "weird" non-German/French accented chars?) accented char, by just entering the respective base char, is quite ingenious (and new info for me, thank you!), and I think it can be of tremendous help IF it works "over" all possible file formats (but I so much doubt this!), and without fault, just compare with File Locator's "handling" (i.e. in fact mis-treating) accented chars even in simple .rtf files (explained in the outliner thread) - thus, if X1 found (sic, I don't dare say "finds") all these hits, by simply entering "relevement", for finding "relèvement" (which could, please note, have been wrongly written rélèvement" in some third-party source text within your "database" / file-system-based data repository, which detail would make you would not find it by entering the correct wording), this would be a very strong argument for using X1, and you clearly should not undervalue this feature, especially since you're a Continental and by this will probably have stored an enormous amount of text bodies containing accented chars, and which rather often will have accent errors within those original texts.

X1 again, a traditional problem of X1 not treated here: What about its handling of OL (Outlook) data? Not only that ancient X1 versions did not treat such data well, but far worse, X1 was deemed, by some commentators, to damage OL files, which of course would be perfectly inacceptable. What about this? I can't trial (neither buy, which I would have done, otherwise) the current X1 version, with my XP Win version, and it might be this obvious X1-vs.-OL problem has been resolved in the meantime (but even then, the question would remain which OL versions would possibly be affected even then? X1-current vs. OL-current possibly ok, but X1-current vs. OL-ancient-versions =?!). I understand that few people would be sufficiently motivated to trial this upon their real data, but then, better trial this, with let's say a replication of your current data, put onto an alternative pc, instead of runningg the risk that even X1-current will damage any OL data on your running system, don't you think so? (And then, thankfully, share your hopeful all-clear signal, or then, your warnings, in case - which would of course be a step further, not necessarily included within your first step of verifying...)

Innuendo on X1 vs. the rest, and in particular dtSearch:

"X1 - Far from perfect, but the absolute best if you use the criteria above as a guideline. Sadly, it seems they are very aware of being the best and have priced their product accordingly. Very expensive...just expensive enough to put it over the line of insulting. If you want the best, you and your wallet will be oh so painfully aware that you are paying for the best."

"dtSearch - This is a solution geared towards corporations and the cold UI and barely there acceptable list of features make this an unappetizing choice for home users. I would wager they make their bones by providing lucrative support plans and willingness to accept company purchase orders. There are more capable, less expensive, more efficient options available."

This cannot stay uncommented since it's obviously wrong in some respects, from my own trialling both; of course, if X1 has got some advantages (beyond the GUI, which indeed is much better, but then, some macroing for dtSearch could probably prevent some premature decision like jity2's one: "In fact after watching some videos about it, I won't try it because I don't use regex for searching keywords, and because the interface seems not very enough user friendly (I don't want to click many times just to do a keyword search !)."), please tell us!

First of all, I can confirm that both developers have (competent) staff (i.e. no comparison with the usual "either it's the developer himself, or some incompetent (since not trained, not informed, not even half-way correctly paid "Indian"") that is really and VERY helpful, in giving information, and in discussing features, or even lack of features, both X1 and dtSearch people are professional and congenial, and if I say dtSearch staff is even "better" than X1 staff, this, while being true, is not to denigrate X1 staff: we're discussing just different degrees of excellence here. (Now compare with Copernic.)

This being said, X1 seems to be visually-brilliant sw for standard applics, whilst dtSearch FINDS IT ALL. In fact, when trialling, I did not encounter any exotic file format from which I wasn't able to get the relevant hits, whilst in X1, if it was not in their (quite standard file format) list, it was not indexed, and thus was not found: It's as simple as that. (Remember the forensic objectives of dtSearch, but it's exactly this additional purpose of it that makes it capable of searching lots of even quite widespread file formats where most other (index-based) desktop search tools fail.

Also, allow for a brief divagation into askSam country: The reason some people cling to it, is the rarity of full-text "db's" able to find numerics. Okay, okay, any search tool can find "386", be it as part of a "string", or even as a "word" (i.e. as a number, or as part of a number), but what about "between 350 and 400"? Okay, okay, you can try (and even succeed, in part), with regex (= again, dtSearch instead of X1). But askSam does this, and similar, with "pseudo-fields", and normally, for such tasks, you need "real" db's for this, and as we all know, for most text-heavy data, people prefer text-based sw, instead of putting it all into relational db's. As you also know, there are some SQLite/other-db-based 2-pane outliners / basic IMS' that have got additional "columns" in order to get numeric data into, but that's not the same (, and even within there, searching for numeric data RANGES is far from evident).

Now that's for numeric ranges in db's, and now look into dtSearch's possibilities of identifying numeric ranges in pseudo-fields in "full text", similar to askSam, and you will see the incredible (and obviously, again, regex-driven) power of dtSearch.

Thus, dear Innuendo, your X1 being "the absolute best" is perfectly unsustainable, but it's in order to inform you better that I post this, and not at all in order to insinuate you had known better whilst writing the above.

____________________

Re ntfs file numbers:

jity2 in the above DC thread: "With CDS V3.6 size of the index was 85 Go with about 2,000,000 files indexed (Note: In one hdd drive I even hit the NTFS limit : too much files to handle !) . It took about 15 days to complete 24/24 7/7." Note: the last info is good to know... ;-(

It's evident 2 million (!) files cannot reach any "NTFS limit" but if you do lots of things completety wrong, and if you persistently left out 3 zeros, it would have been 8.6 (or, with the XP number, 4.3, but nothing near 2.0:)

eVista on

https://social.technet.microsoft.com/Forums/windows/en-US/772dbf1a-536c-47d3-8a8d-f773c90b8a5e/maximum-number-of-files-allowed-to-exist-per-ntfs-volume-under-windows-vista-32bit-home-premium?forum=itprovistaapps :

"In short, the absolute limit on the number of files per NTFS volume seems to be 2 at the 32nd power minus 1*, but this would require 512 byte sectors and a maximum file size limit of one file per sector. Therefore, in practice, one has to calculate a realistic average file size and then apply these principles to that file size."

Note: That would be a little less than 4.3 (i.e. 2power32-1) billion files (for Continentals: 4,3 Milliarden/milliards/etc.), for XP, whilst it's 2power64-1 for Vista on, i.e. slightly less than 8.6 billion files.

EDIT: OF COURSE THAT IS NOT TRUE: The number you get everywhere is 2power32 = slightly less than 4.3 billion files, and I read that's for XP, whilst from Vista on, it would be double of that, which would make it a little less than 8.6 indeed (I cannot confirm this of course), and that would then be 2power33, not 64 (I obviously got lead astray by Win32/64 (which probably is behind that doubling though)).

No need to list all the google finds, just let me say that with "ntfs file number" you'll get the results you need, incl. wikipedia, MS...

But then, special mention to http://stackoverflow.com/questions/197162/ntfs-performance-and-large-volumes-of-files-and-directories

with an absolutely brilliant "best answer", and then also lots of valuable details further down that page.

I think this last link will give you plenty of ideas how to better organize your stuff, but anyway, no search tool whatsoever should choke by some "2,000,000 limit", ntfs or otherwise.

Pages: [1] 2 3 4 5next
Go to full version