Wednesday, 5 June 2013

My ideas on Character Recognition

Character Recognition Summary written 06/03/13 10:26 [Wednesday]
Since my teenage years I have been interested in visual perception: how humans (and animals) see things and recognise them and thereby negotiate the world. Rather than go into the entire history of this interest of mine, though, my intention here is to summarise the programming work I have done over recent years to implement ideas I have had on how particularly humans view the world using their eyes.
In 2006 the computer I owned was capable, for about the first time since I had owned a computer, of doing in a reasonable time the lengthy computation I needed to compare printed characters (that is characters with exemplars to try to identify what the characters were, from the alphabet) according to a formula I had thought up for a measure of similarity. I published on my website a description of my ideas with some specific results from what I may call experimentation. Towards the end of 2006 I was grappling with the problem of printed characters running together through over-inking instead of standing separately, and thinking in terms of the aspect ratio characters have on average as a method for trying to separate individual characters.
This led in a natural course to my trying in 2007 to vary the greyscale threshold for distinguishing black from white in such a way as to separate out characters one from the next. Instead of looking for seeming characters with an aspect ratio in a particular range though I had the idea of measuring how fragments of black emerged with a raising of the greyscale threshold and preferring the range of threshold where the fragments appeared most stable. This had the advantages that the estimation could be done locally for neighbourhoods within the scan (so that shadows across part of a document would not throw the estimation out) and that the method could be applied to general images and not just pictures of objects within a known range of aspect ratio.
By 2008 instead of using the raw greyscale for each pixel I was doing what I call a ‘blackdensity’ computation so that black pixels locally to a given pixel will increase the measure of blackness at that pixel. By taking a local average in this way (but an average where closer pixels have more weight) mistaken measurements from the scan at particular pixels are evened out. One thing coming out of this methodology was that peaks of blackdensity (‘saliences’) could be counted and the way the count varied as resolution varied could be observed. This led to the hope that there would be a ‘natural’ scale of distances (that is, resolution) for a given pattern imaged so that the same object seen from different distances - giving the same pattern of saliences but scaled differently - might still be recognised.
From 2009 to 2011 because of the condition of mind I was in I got bogged down in too much detail and the work on ‘Visual field analysis’ was in abeyance, except for the general idea emerging in my mind of using a measure of ‘busyness’ of fragments to indicate how useful was the information contained in the pattern. The correct way to measure information in a greyscale pattern I now think is to compute what I call the clustermeasure. This has a formula very like the similaritymeasure which I was using in 2006, except that similaritymeasure is for two different patterns being compared. One thing I did achieve in 2009 was a very lucid explanation of clustermeasure and its additive simplicity as new clusters are added.

Wednesday, 3 April 2013

Absolutely correct


03/04/13 03:13 [Wednesday]
I woke up about 2.30 AM and I can’t say this is the reason I was awake and couldn’t get back to sleep (that had more to do with the headache I had and discomfort in my tummy) but running constantly through my mind was the phrase ‘in-depth profile’. This phrase was used in a TV ad for a series of magazines when I was a teenager, and is troublesome in that a profile is an outline so how can it have depth? I was trying to think whether any meaning could be ascribed to ‘in-depth profile’ as I was waking up this morning.
Since I have got up I have been thinking about a sequence in The Big Bang Theory where Sheldon says one cannot be ‘more wrong’ because wrong is absolute: on a particular question one is either wrong or not wrong. And I have been thinking of other similar phrases used nowadays less strictly than really they ought to be: ‘more essential’ and ‘very key’.
The person replying to Sheldon on that matter in The Big Bang Theory (the comic store owner, I think) said it was a little wrong to say that a tomato is a vegetable but a lot wrong to say something else which I can’t recall but which was unarguably wrong or I should say more commonly declared wrong. Without getting too much into the question of the nature of truth and the nature of error, it is ‘unarguably wrong’ that a tomato is a vegetable but it is quite commonly thought to be right not wrong.
So for some qualities which are actually binary - either the case or not the case - there are associated measures of degree which seem natural. To say ‘more pregnant’ means in fact ‘further along in the pregnancy’. And for the case of wrongness there are as it were votes which could be taken showing what proportion of people realise (or ‘believe’) that the thing is wrong and what proportion believe it to be right.
If something is essential to something it is a necessary condition for it. Guessing what ‘more essential’ might mean (without having to hand a specific example) I suppose it means ‘more likely to fail if this condition is absent’, in which case the condition is not in truth essential but is simply important. Or possibly we can resort to voting again: perhaps we do not know whether this condition is essential but a certain large proportion of people believe the thing would fail without the condition.
I’m not quite sure what ‘key’ means not having thought through the etymology, but in ‘very key’ it means ‘important’ or ‘significant’. ‘Very key’ sounds wrong to me but it is quite a common phrase.
People like Sheldon Cooper (and Mrs Thatcher as she used to be) in coming to conclusions or decisions are very sure of them. Personally I think this is related to the way dopamine fires up in their central nervous systems, that is when dopaminergic neurons are excited they are very excited very suddenly and when they cease to be excited they cease suddenly (and this of course is related to schizophrenia). Commonly nowadays though individuals are not trusted: managers do not take decisions based on what they know to be the case but rather a questionnaire is sent out to determine what people on average believe to be the case. Hence I can understand the more frequent use nowadays of phrases like ‘more wrong’ which depend on the notion of what people on average believe instead of what is in truth the case.
Addendum
According to dictionary.reference.com (the online OED via the Dudley Council website being unavailable) key (as an adjective) means chief; major; important; essential; fundamental; pivotal. (I recommend also www.merriam-webster.com which I have discovered through the unavailability of the OED.) I should think the prior meaning of key is essential as a key is a sine qua non for getting through a door (or something else locked), and hence my doubts about ‘very key’ which is like saying ‘very essential’.
Nowadays the idea that anything is absolute is eroded, and even in physics the belief has become prevalent that there is not any absolute truth but only theories which come and go and for a while explain things quite well.